[ 555.041232] env[62692]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=62692) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 555.041641] env[62692]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=62692) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 555.041641] env[62692]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=62692) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 555.041959] env[62692]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 555.137265] env[62692]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=62692) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 555.147402] env[62692]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=62692) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 555.746555] env[62692]: INFO nova.virt.driver [None req-100ad448-b1a1-499a-b636-f7b7d0f84929 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 555.816811] env[62692]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.817051] env[62692]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.817119] env[62692]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=62692) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 558.904364] env[62692]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-b2b9941e-0521-4a17-b60c-977b65e977f1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.920263] env[62692]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=62692) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 558.920402] env[62692]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-7ced425d-d894-4f4d-9387-25eeedeb2965 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.946094] env[62692]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 4d774. [ 558.946212] env[62692]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.129s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 558.946752] env[62692]: INFO nova.virt.vmwareapi.driver [None req-100ad448-b1a1-499a-b636-f7b7d0f84929 None None] VMware vCenter version: 7.0.3 [ 558.950088] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b913b4ff-1949-4893-9668-98b5807aba78 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.971614] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eb33ef5-6e1c-4c63-a09c-217395924984 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.977382] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22cc269a-346c-4ad0-bf3b-c1f9594e9bb9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.983822] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62d55841-9b50-461f-97b2-5d92acb66b42 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.996571] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33fc8de8-74fb-408f-8e2d-ad8bda719623 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.002212] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b7d4790-6fcf-40f0-8145-960d82475903 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.031457] env[62692]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-4b6e76bf-24e3-4179-bf6a-e4245f6bd36c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.036215] env[62692]: DEBUG nova.virt.vmwareapi.driver [None req-100ad448-b1a1-499a-b636-f7b7d0f84929 None None] Extension org.openstack.compute already exists. {{(pid=62692) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:224}} [ 559.038806] env[62692]: INFO nova.compute.provider_config [None req-100ad448-b1a1-499a-b636-f7b7d0f84929 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 559.542749] env[62692]: DEBUG nova.context [None req-100ad448-b1a1-499a-b636-f7b7d0f84929 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),28470fb0-2760-4a04-8314-f7088f0543bd(cell1) {{(pid=62692) load_cells /opt/stack/nova/nova/context.py:464}} [ 559.545503] env[62692]: DEBUG oslo_concurrency.lockutils [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.545734] env[62692]: DEBUG oslo_concurrency.lockutils [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.546523] env[62692]: DEBUG oslo_concurrency.lockutils [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 559.546992] env[62692]: DEBUG oslo_concurrency.lockutils [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] Acquiring lock "28470fb0-2760-4a04-8314-f7088f0543bd" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.547202] env[62692]: DEBUG oslo_concurrency.lockutils [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] Lock "28470fb0-2760-4a04-8314-f7088f0543bd" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.548223] env[62692]: DEBUG oslo_concurrency.lockutils [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] Lock "28470fb0-2760-4a04-8314-f7088f0543bd" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 559.568078] env[62692]: INFO dbcounter [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] Registered counter for database nova_cell0 [ 559.576194] env[62692]: INFO dbcounter [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] Registered counter for database nova_cell1 [ 559.579378] env[62692]: DEBUG oslo_db.sqlalchemy.engines [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62692) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 559.579802] env[62692]: DEBUG oslo_db.sqlalchemy.engines [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62692) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 559.584557] env[62692]: ERROR nova.db.main.api [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 559.584557] env[62692]: result = function(*args, **kwargs) [ 559.584557] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 559.584557] env[62692]: return func(*args, **kwargs) [ 559.584557] env[62692]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 559.584557] env[62692]: result = fn(*args, **kwargs) [ 559.584557] env[62692]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 559.584557] env[62692]: return f(*args, **kwargs) [ 559.584557] env[62692]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 559.584557] env[62692]: return db.service_get_minimum_version(context, binaries) [ 559.584557] env[62692]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 559.584557] env[62692]: _check_db_access() [ 559.584557] env[62692]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 559.584557] env[62692]: stacktrace = ''.join(traceback.format_stack()) [ 559.584557] env[62692]: [ 559.585576] env[62692]: ERROR nova.db.main.api [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 559.585576] env[62692]: result = function(*args, **kwargs) [ 559.585576] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 559.585576] env[62692]: return func(*args, **kwargs) [ 559.585576] env[62692]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 559.585576] env[62692]: result = fn(*args, **kwargs) [ 559.585576] env[62692]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 559.585576] env[62692]: return f(*args, **kwargs) [ 559.585576] env[62692]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 559.585576] env[62692]: return db.service_get_minimum_version(context, binaries) [ 559.585576] env[62692]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 559.585576] env[62692]: _check_db_access() [ 559.585576] env[62692]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 559.585576] env[62692]: stacktrace = ''.join(traceback.format_stack()) [ 559.585576] env[62692]: [ 559.585940] env[62692]: WARNING nova.objects.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 559.586106] env[62692]: WARNING nova.objects.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] Failed to get minimum service version for cell 28470fb0-2760-4a04-8314-f7088f0543bd [ 559.586547] env[62692]: DEBUG oslo_concurrency.lockutils [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] Acquiring lock "singleton_lock" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 559.586719] env[62692]: DEBUG oslo_concurrency.lockutils [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] Acquired lock "singleton_lock" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.586964] env[62692]: DEBUG oslo_concurrency.lockutils [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] Releasing lock "singleton_lock" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.587314] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] Full set of CONF: {{(pid=62692) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 559.587459] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ******************************************************************************** {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 559.587588] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] Configuration options gathered from: {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 559.587728] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 559.587912] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 559.588054] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ================================================================================ {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 559.588265] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] allow_resize_to_same_host = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.588439] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] arq_binding_timeout = 300 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.588604] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] backdoor_port = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.588749] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] backdoor_socket = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.588917] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] block_device_allocate_retries = 60 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.589093] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] block_device_allocate_retries_interval = 3 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.589268] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cert = self.pem {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.589435] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.589608] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] compute_monitors = [] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.589774] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] config_dir = [] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.589943] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] config_drive_format = iso9660 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.590091] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.590255] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] config_source = [] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.590420] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] console_host = devstack {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.590584] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] control_exchange = nova {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.590742] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cpu_allocation_ratio = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.590900] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] daemon = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.591106] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] debug = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.591241] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] default_access_ip_network_name = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.591410] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] default_availability_zone = nova {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.591563] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] default_ephemeral_format = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.591767] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] default_green_pool_size = 1000 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.592047] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.592256] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] default_schedule_zone = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.592449] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] disk_allocation_ratio = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.592631] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] enable_new_services = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.592815] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] enabled_apis = ['osapi_compute'] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.592978] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] enabled_ssl_apis = [] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.593157] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] flat_injected = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.593318] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] force_config_drive = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.593524] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] force_raw_images = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.593743] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] graceful_shutdown_timeout = 5 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.593958] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] heal_instance_info_cache_interval = 60 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.594254] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] host = cpu-1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.594461] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.594672] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] initial_disk_allocation_ratio = 1.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.594878] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] initial_ram_allocation_ratio = 1.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.595135] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.595311] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] instance_build_timeout = 0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.595506] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] instance_delete_interval = 300 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.595681] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] instance_format = [instance: %(uuid)s] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.595850] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] instance_name_template = instance-%08x {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.596019] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] instance_usage_audit = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.596194] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] instance_usage_audit_period = month {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.596363] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.596561] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] instances_path = /opt/stack/data/nova/instances {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.596744] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] internal_service_availability_zone = internal {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.596906] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] key = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.597078] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] live_migration_retry_count = 30 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.597250] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] log_color = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.597415] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] log_config_append = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.597580] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.597741] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] log_dir = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.597896] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] log_file = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.598032] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] log_options = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.598202] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] log_rotate_interval = 1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.598372] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] log_rotate_interval_type = days {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.598534] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] log_rotation_type = none {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.598662] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.598786] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.598951] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.599128] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.599256] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.599418] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] long_rpc_timeout = 1800 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.599579] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] max_concurrent_builds = 10 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.599740] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] max_concurrent_live_migrations = 1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.599894] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] max_concurrent_snapshots = 5 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.600063] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] max_local_block_devices = 3 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.600226] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] max_logfile_count = 30 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.600384] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] max_logfile_size_mb = 200 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.600542] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] maximum_instance_delete_attempts = 5 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.600706] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] metadata_listen = 0.0.0.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.600870] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] metadata_listen_port = 8775 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.601050] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] metadata_workers = 2 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.601214] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] migrate_max_retries = -1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.601379] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] mkisofs_cmd = genisoimage {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.601585] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] my_block_storage_ip = 10.180.1.21 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.601716] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] my_ip = 10.180.1.21 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.601880] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] network_allocate_retries = 0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.602069] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.602241] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] osapi_compute_listen = 0.0.0.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.602407] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] osapi_compute_listen_port = 8774 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.602576] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] osapi_compute_unique_server_name_scope = {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.602745] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] osapi_compute_workers = 2 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.602912] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] password_length = 12 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.603086] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] periodic_enable = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.603252] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] periodic_fuzzy_delay = 60 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.603423] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] pointer_model = usbtablet {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.603591] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] preallocate_images = none {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.603751] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] publish_errors = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.603878] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] pybasedir = /opt/stack/nova {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.604045] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ram_allocation_ratio = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.604210] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] rate_limit_burst = 0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.604379] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] rate_limit_except_level = CRITICAL {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.604540] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] rate_limit_interval = 0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.604740] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] reboot_timeout = 0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.604916] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] reclaim_instance_interval = 0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.605091] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] record = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.605268] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] reimage_timeout_per_gb = 60 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.605466] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] report_interval = 120 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.605644] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] rescue_timeout = 0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.605808] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] reserved_host_cpus = 0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.606016] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] reserved_host_disk_mb = 0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.606236] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] reserved_host_memory_mb = 512 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.606403] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] reserved_huge_pages = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.606569] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] resize_confirm_window = 0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.606756] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] resize_fs_using_block_device = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.606931] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] resume_guests_state_on_host_boot = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.607120] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.607289] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] rpc_response_timeout = 60 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.607452] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] run_external_periodic_tasks = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.607623] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] running_deleted_instance_action = reap {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.607787] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] running_deleted_instance_poll_interval = 1800 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.607947] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] running_deleted_instance_timeout = 0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.608121] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] scheduler_instance_sync_interval = 120 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.608293] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] service_down_time = 720 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.608463] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] servicegroup_driver = db {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.608620] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] shell_completion = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.608780] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] shelved_offload_time = 0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.608942] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] shelved_poll_interval = 3600 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.609121] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] shutdown_timeout = 0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.609286] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] source_is_ipv6 = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.609445] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ssl_only = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.609690] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.609858] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] sync_power_state_interval = 600 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.610031] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] sync_power_state_pool_size = 1000 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.610204] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] syslog_log_facility = LOG_USER {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.610363] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] tempdir = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.610528] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] timeout_nbd = 10 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.610693] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] transport_url = **** {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.610855] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] update_resources_interval = 0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.611023] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] use_cow_images = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.611189] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] use_eventlog = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.611349] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] use_journal = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.611511] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] use_json = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.611670] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] use_rootwrap_daemon = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.611828] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] use_stderr = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.611986] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] use_syslog = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.612155] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vcpu_pin_set = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.612324] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vif_plugging_is_fatal = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.612493] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vif_plugging_timeout = 300 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.612661] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] virt_mkfs = [] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.612822] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] volume_usage_poll_interval = 0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.612984] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] watch_log_file = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.613166] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] web = /usr/share/spice-html5 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 559.613351] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_concurrency.disable_process_locking = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.613640] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.613823] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.613993] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.614184] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.614362] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.614534] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.614719] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api.auth_strategy = keystone {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.614889] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api.compute_link_prefix = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.615080] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.615261] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api.dhcp_domain = novalocal {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.615463] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api.enable_instance_password = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.615643] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api.glance_link_prefix = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.615813] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.615986] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.616166] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api.instance_list_per_project_cells = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.616335] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api.list_records_by_skipping_down_cells = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.616536] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api.local_metadata_per_cell = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.616736] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api.max_limit = 1000 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.616912] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api.metadata_cache_expiration = 15 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.617105] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api.neutron_default_tenant_id = default {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.617352] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api.response_validation = warn {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.617559] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api.use_neutron_default_nets = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.617740] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.617911] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.618103] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.618288] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.618465] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api.vendordata_dynamic_targets = [] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.618633] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api.vendordata_jsonfile_path = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.618862] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.619084] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.backend = dogpile.cache.memcached {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.619263] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.backend_argument = **** {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.619438] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.config_prefix = cache.oslo {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.619611] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.dead_timeout = 60.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.619776] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.debug_cache_backend = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.619942] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.enable_retry_client = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.620120] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.enable_socket_keepalive = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.620299] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.enabled = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.620466] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.enforce_fips_mode = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.620637] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.expiration_time = 600 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.620804] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.hashclient_retry_attempts = 2 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.620973] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.hashclient_retry_delay = 1.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.621154] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.memcache_dead_retry = 300 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.621316] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.memcache_password = **** {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.621486] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.621653] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.621820] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.memcache_pool_maxsize = 10 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.621986] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.622164] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.memcache_sasl_enabled = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.622349] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.622524] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.memcache_socket_timeout = 1.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.622689] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.memcache_username = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.622859] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.proxies = [] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.623038] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.redis_db = 0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.623207] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.redis_password = **** {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.623381] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.redis_sentinel_service_name = mymaster {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.623557] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.623726] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.redis_server = localhost:6379 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.623894] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.redis_socket_timeout = 1.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.624068] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.redis_username = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.624237] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.retry_attempts = 2 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.624404] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.retry_delay = 0.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.624568] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.socket_keepalive_count = 1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.624734] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.socket_keepalive_idle = 1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.624897] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.socket_keepalive_interval = 1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.625068] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.tls_allowed_ciphers = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.625230] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.tls_cafile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.625414] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.tls_certfile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.625597] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.tls_enabled = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.625762] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cache.tls_keyfile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.625945] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cinder.auth_section = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.626144] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cinder.auth_type = password {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.626311] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cinder.cafile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.626534] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cinder.catalog_info = volumev3::publicURL {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.626712] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cinder.certfile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.626883] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cinder.collect_timing = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.627061] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cinder.cross_az_attach = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.627232] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cinder.debug = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.627398] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cinder.endpoint_template = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.627567] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cinder.http_retries = 3 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.627732] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cinder.insecure = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.627891] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cinder.keyfile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.628084] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cinder.os_region_name = RegionOne {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.628257] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cinder.split_loggers = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.628422] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cinder.timeout = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.628625] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.628806] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] compute.cpu_dedicated_set = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.628969] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] compute.cpu_shared_set = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.629154] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] compute.image_type_exclude_list = [] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.629324] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.629496] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] compute.max_concurrent_disk_ops = 0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.629661] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] compute.max_disk_devices_to_attach = -1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.629827] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.630015] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.630181] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] compute.resource_provider_association_refresh = 300 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.630415] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.630618] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] compute.shutdown_retry_interval = 10 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.630835] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.631049] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] conductor.workers = 2 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.631242] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] console.allowed_origins = [] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.631410] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] console.ssl_ciphers = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.631588] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] console.ssl_minimum_version = default {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.631759] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] consoleauth.enforce_session_timeout = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.631931] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] consoleauth.token_ttl = 600 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.632120] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cyborg.cafile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.632284] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cyborg.certfile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.632450] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cyborg.collect_timing = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.632613] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cyborg.connect_retries = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.632773] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cyborg.connect_retry_delay = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.632932] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cyborg.endpoint_override = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.633110] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cyborg.insecure = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.633272] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cyborg.keyfile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.633434] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cyborg.max_version = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.633603] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cyborg.min_version = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.633754] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cyborg.region_name = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.633911] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cyborg.retriable_status_codes = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.634079] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cyborg.service_name = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.634251] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cyborg.service_type = accelerator {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.634415] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cyborg.split_loggers = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.634579] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cyborg.status_code_retries = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.634739] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cyborg.status_code_retry_delay = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.634898] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cyborg.timeout = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.635093] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.635259] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] cyborg.version = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.635467] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] database.backend = sqlalchemy {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.635651] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] database.connection = **** {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.635823] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] database.connection_debug = 0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.635996] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] database.connection_parameters = {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.636201] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] database.connection_recycle_time = 3600 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.636433] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] database.connection_trace = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.636644] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] database.db_inc_retry_interval = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.636823] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] database.db_max_retries = 20 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.636992] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] database.db_max_retry_interval = 10 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.637176] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] database.db_retry_interval = 1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.637342] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] database.max_overflow = 50 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.637508] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] database.max_pool_size = 5 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.637674] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] database.max_retries = 10 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.637844] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.638013] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] database.mysql_wsrep_sync_wait = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.638179] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] database.pool_timeout = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.638344] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] database.retry_interval = 10 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.638504] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] database.slave_connection = **** {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.638667] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] database.sqlite_synchronous = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.638826] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] database.use_db_reconnect = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.639021] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api_database.backend = sqlalchemy {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.639191] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api_database.connection = **** {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.639359] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api_database.connection_debug = 0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.639532] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api_database.connection_parameters = {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.639858] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api_database.connection_recycle_time = 3600 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.639858] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api_database.connection_trace = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.640018] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api_database.db_inc_retry_interval = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.640190] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api_database.db_max_retries = 20 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.640355] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api_database.db_max_retry_interval = 10 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.640521] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api_database.db_retry_interval = 1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.640720] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api_database.max_overflow = 50 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.640889] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api_database.max_pool_size = 5 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.641101] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api_database.max_retries = 10 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.641289] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.641456] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.641620] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api_database.pool_timeout = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.641785] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api_database.retry_interval = 10 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.641938] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api_database.slave_connection = **** {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.642128] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] api_database.sqlite_synchronous = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.642306] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] devices.enabled_mdev_types = [] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.642520] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.642702] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ephemeral_storage_encryption.default_format = luks {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.642916] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ephemeral_storage_encryption.enabled = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.643118] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.643299] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.api_servers = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.643469] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.cafile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.643635] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.certfile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.643801] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.collect_timing = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.643961] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.connect_retries = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.644136] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.connect_retry_delay = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.644299] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.debug = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.644465] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.default_trusted_certificate_ids = [] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.644630] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.enable_certificate_validation = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.644793] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.enable_rbd_download = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.644950] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.endpoint_override = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.645130] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.insecure = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.645295] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.keyfile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.645487] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.max_version = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.645658] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.min_version = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.645825] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.num_retries = 3 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.645997] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.rbd_ceph_conf = {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.646179] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.rbd_connect_timeout = 5 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.646361] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.rbd_pool = {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.646618] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.rbd_user = {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.646817] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.region_name = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.646985] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.retriable_status_codes = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.647166] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.service_name = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.647342] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.service_type = image {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.647509] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.split_loggers = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.647674] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.status_code_retries = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.647834] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.status_code_retry_delay = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.647995] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.timeout = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.648201] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.648373] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.verify_glance_signatures = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.648536] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] glance.version = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.648705] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] guestfs.debug = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.648876] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] mks.enabled = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.649244] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.649440] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] image_cache.manager_interval = 2400 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.649614] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] image_cache.precache_concurrency = 1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.649786] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] image_cache.remove_unused_base_images = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.649957] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.650141] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.650319] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] image_cache.subdirectory_name = _base {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.650496] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.api_max_retries = 60 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.650662] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.api_retry_interval = 2 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.650822] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.auth_section = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.650986] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.auth_type = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.651161] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.cafile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.651321] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.certfile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.651484] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.collect_timing = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.651652] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.conductor_group = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.651813] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.connect_retries = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.651975] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.connect_retry_delay = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.652149] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.endpoint_override = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.652311] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.insecure = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.652473] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.keyfile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.652655] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.max_version = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.652843] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.min_version = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.653034] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.peer_list = [] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.653201] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.region_name = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.653364] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.retriable_status_codes = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.653529] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.serial_console_state_timeout = 10 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.653689] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.service_name = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.653880] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.service_type = baremetal {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.654021] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.shard = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.654187] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.split_loggers = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.654361] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.status_code_retries = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.654556] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.status_code_retry_delay = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.654727] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.timeout = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.654925] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.655126] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ironic.version = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.655318] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.655520] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] key_manager.fixed_key = **** {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.655713] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.655877] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] barbican.barbican_api_version = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.656050] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] barbican.barbican_endpoint = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.656227] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] barbican.barbican_endpoint_type = public {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.656387] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] barbican.barbican_region_name = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.656549] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] barbican.cafile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.656709] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] barbican.certfile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.656873] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] barbican.collect_timing = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.657047] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] barbican.insecure = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.657210] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] barbican.keyfile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.657376] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] barbican.number_of_retries = 60 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.657541] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] barbican.retry_delay = 1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.657705] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] barbican.send_service_user_token = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.657867] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] barbican.split_loggers = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.658034] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] barbican.timeout = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.658203] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] barbican.verify_ssl = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.658362] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] barbican.verify_ssl_path = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.658570] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] barbican_service_user.auth_section = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.658744] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] barbican_service_user.auth_type = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.658904] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] barbican_service_user.cafile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.659076] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] barbican_service_user.certfile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.659246] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] barbican_service_user.collect_timing = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.659410] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] barbican_service_user.insecure = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.659570] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] barbican_service_user.keyfile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.659734] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] barbican_service_user.split_loggers = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.659894] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] barbican_service_user.timeout = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.660076] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vault.approle_role_id = **** {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.660238] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vault.approle_secret_id = **** {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.660407] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vault.kv_mountpoint = secret {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.660571] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vault.kv_path = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.660736] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vault.kv_version = 2 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.660897] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vault.namespace = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.661066] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vault.root_token_id = **** {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.661228] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vault.ssl_ca_crt_file = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.661399] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vault.timeout = 60.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.661563] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vault.use_ssl = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.661733] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.661907] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] keystone.auth_section = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.662083] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] keystone.auth_type = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.662243] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] keystone.cafile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.662403] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] keystone.certfile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.662568] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] keystone.collect_timing = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.662725] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] keystone.connect_retries = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.662883] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] keystone.connect_retry_delay = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.663052] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] keystone.endpoint_override = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.663215] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] keystone.insecure = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.663373] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] keystone.keyfile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.663530] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] keystone.max_version = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.663686] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] keystone.min_version = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.663843] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] keystone.region_name = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.663996] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] keystone.retriable_status_codes = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.664166] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] keystone.service_name = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.664338] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] keystone.service_type = identity {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.664502] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] keystone.split_loggers = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.664689] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] keystone.status_code_retries = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.664880] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] keystone.status_code_retry_delay = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.665055] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] keystone.timeout = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.665241] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.665428] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] keystone.version = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.665647] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.connection_uri = {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.665814] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.cpu_mode = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.665984] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.cpu_model_extra_flags = [] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.666173] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.cpu_models = [] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.666352] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.cpu_power_governor_high = performance {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.666559] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.cpu_power_governor_low = powersave {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.666729] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.cpu_power_management = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.666900] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.667115] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.device_detach_attempts = 8 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.667290] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.device_detach_timeout = 20 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.667458] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.disk_cachemodes = [] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.667623] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.disk_prefix = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.667789] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.enabled_perf_events = [] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.667951] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.file_backed_memory = 0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.668130] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.gid_maps = [] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.668292] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.hw_disk_discard = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.668453] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.hw_machine_type = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.668626] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.images_rbd_ceph_conf = {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.668793] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.668959] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.669144] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.images_rbd_glance_store_name = {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.669322] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.images_rbd_pool = rbd {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.669496] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.images_type = default {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.669659] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.images_volume_group = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.669824] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.inject_key = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.669989] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.inject_partition = -2 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.670167] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.inject_password = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.670333] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.iscsi_iface = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.670530] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.iser_use_multipath = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.670728] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.live_migration_bandwidth = 0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.670899] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.671078] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.live_migration_downtime = 500 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.671247] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.671412] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.671575] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.live_migration_inbound_addr = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.671741] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.671904] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.live_migration_permit_post_copy = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.672074] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.live_migration_scheme = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.672252] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.live_migration_timeout_action = abort {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.672423] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.live_migration_tunnelled = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.672588] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.live_migration_uri = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.672756] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.live_migration_with_native_tls = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.672921] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.max_queues = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.673098] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.673331] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.673499] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.nfs_mount_options = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.674640] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.674837] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.675028] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.num_iser_scan_tries = 5 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.675196] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.num_memory_encrypted_guests = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.675376] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.675560] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.num_pcie_ports = 0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.675735] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.num_volume_scan_tries = 5 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.675907] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.pmem_namespaces = [] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.676081] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.quobyte_client_cfg = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.676380] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.676573] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.rbd_connect_timeout = 5 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.676745] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.676911] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.677087] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.rbd_secret_uuid = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.677252] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.rbd_user = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.677418] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.677592] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.remote_filesystem_transport = ssh {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.677754] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.rescue_image_id = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.677913] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.rescue_kernel_id = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.678083] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.rescue_ramdisk_id = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.678255] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.678437] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.rx_queue_size = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.678636] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.smbfs_mount_options = {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.678915] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.679103] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.snapshot_compression = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.679273] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.snapshot_image_format = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.679495] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.679666] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.sparse_logical_volumes = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.679832] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.swtpm_enabled = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.680011] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.swtpm_group = tss {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.680188] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.swtpm_user = tss {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.680358] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.sysinfo_serial = unique {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.680521] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.tb_cache_size = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.680680] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.tx_queue_size = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.680846] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.uid_maps = [] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.681017] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.use_virtio_for_bridges = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.681192] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.virt_type = kvm {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.681363] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.volume_clear = zero {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.681532] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.volume_clear_size = 0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.681700] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.volume_use_multipath = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.681861] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.vzstorage_cache_path = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.682048] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.682225] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.vzstorage_mount_group = qemu {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.682392] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.vzstorage_mount_opts = [] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.682589] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.682883] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.683079] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.vzstorage_mount_user = stack {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.683255] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.683436] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.auth_section = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.683615] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.auth_type = password {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.683780] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.cafile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.683944] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.certfile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.684154] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.collect_timing = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.684289] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.connect_retries = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.684454] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.connect_retry_delay = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.684627] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.default_floating_pool = public {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.684790] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.endpoint_override = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.684957] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.extension_sync_interval = 600 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.685141] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.http_retries = 3 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.685307] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.insecure = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.685502] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.keyfile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.685674] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.max_version = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.685849] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.686022] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.min_version = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.686197] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.ovs_bridge = br-int {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.686386] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.physnets = [] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.686624] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.region_name = RegionOne {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.686804] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.retriable_status_codes = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.686984] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.service_metadata_proxy = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.687167] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.service_name = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.687342] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.service_type = network {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.687511] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.split_loggers = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.687676] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.status_code_retries = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.687839] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.status_code_retry_delay = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.688008] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.timeout = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.688202] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.688368] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] neutron.version = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.688554] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] notifications.bdms_in_notifications = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.688736] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] notifications.default_level = INFO {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.688912] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] notifications.notification_format = unversioned {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.689091] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] notifications.notify_on_state_change = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.689273] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.689454] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] pci.alias = [] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.689626] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] pci.device_spec = [] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.689793] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] pci.report_in_placement = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.689969] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.auth_section = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.690159] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.auth_type = password {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.690331] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.690523] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.cafile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.690704] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.certfile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.690871] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.collect_timing = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.691043] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.connect_retries = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.691212] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.connect_retry_delay = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.691373] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.default_domain_id = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.691534] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.default_domain_name = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.691694] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.domain_id = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.691852] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.domain_name = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.692013] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.endpoint_override = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.692183] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.insecure = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.692343] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.keyfile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.692501] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.max_version = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.692661] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.min_version = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.692830] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.password = **** {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.692989] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.project_domain_id = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.693171] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.project_domain_name = Default {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.693340] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.project_id = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.693515] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.project_name = service {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.693685] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.region_name = RegionOne {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.693848] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.retriable_status_codes = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.694012] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.service_name = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.694188] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.service_type = placement {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.694356] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.split_loggers = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.694515] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.status_code_retries = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.694712] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.status_code_retry_delay = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.694886] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.system_scope = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.695061] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.timeout = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.695226] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.trust_id = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.695408] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.user_domain_id = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.695597] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.user_domain_name = Default {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.695763] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.user_id = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.695940] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.username = nova {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.696138] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.696302] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] placement.version = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.696506] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] quota.cores = 20 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.696679] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] quota.count_usage_from_placement = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.696853] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.697035] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] quota.injected_file_content_bytes = 10240 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.697208] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] quota.injected_file_path_length = 255 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.697376] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] quota.injected_files = 5 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.697545] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] quota.instances = 10 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.697714] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] quota.key_pairs = 100 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.697880] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] quota.metadata_items = 128 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.698057] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] quota.ram = 51200 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.698226] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] quota.recheck_quota = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.698396] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] quota.server_group_members = 10 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.698565] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] quota.server_groups = 10 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.698740] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.698906] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.699080] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] scheduler.image_metadata_prefilter = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.699245] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.699415] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] scheduler.max_attempts = 3 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.699581] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] scheduler.max_placement_results = 1000 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.699784] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.700014] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] scheduler.query_placement_for_image_type_support = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.700201] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.700381] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] scheduler.workers = 2 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.700569] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.700744] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.700923] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.701107] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.701280] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.701448] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.701615] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.701809] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.701985] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] filter_scheduler.host_subset_size = 1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.702241] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.702454] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.702635] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.702808] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] filter_scheduler.isolated_hosts = [] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.702982] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] filter_scheduler.isolated_images = [] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.703169] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.703337] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.703504] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.703673] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] filter_scheduler.pci_in_placement = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.703839] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.704010] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.704188] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.704386] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.704519] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.704685] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.704849] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] filter_scheduler.track_instance_changes = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.705042] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.705220] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] metrics.required = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.705411] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] metrics.weight_multiplier = 1.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.705576] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.705750] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] metrics.weight_setting = [] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.706093] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.706277] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] serial_console.enabled = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.706492] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] serial_console.port_range = 10000:20000 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.706682] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.706857] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.707039] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] serial_console.serialproxy_port = 6083 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.707216] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] service_user.auth_section = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.707393] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] service_user.auth_type = password {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.707558] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] service_user.cafile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.707720] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] service_user.certfile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.707885] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] service_user.collect_timing = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.708057] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] service_user.insecure = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.708222] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] service_user.keyfile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.708396] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] service_user.send_service_user_token = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.708565] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] service_user.split_loggers = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.708729] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] service_user.timeout = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.708900] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] spice.agent_enabled = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.709074] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] spice.enabled = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.709405] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.709609] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.709786] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] spice.html5proxy_port = 6082 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.709951] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] spice.image_compression = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.710129] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] spice.jpeg_compression = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.710295] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] spice.playback_compression = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.710471] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] spice.server_listen = 127.0.0.1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.710645] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.710806] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] spice.streaming_mode = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.710968] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] spice.zlib_compression = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.711151] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] upgrade_levels.baseapi = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.711325] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] upgrade_levels.compute = auto {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.711489] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] upgrade_levels.conductor = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.711650] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] upgrade_levels.scheduler = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.711816] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vendordata_dynamic_auth.auth_section = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.711980] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vendordata_dynamic_auth.auth_type = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.712155] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vendordata_dynamic_auth.cafile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.712314] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vendordata_dynamic_auth.certfile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.712479] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.712644] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vendordata_dynamic_auth.insecure = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.712802] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vendordata_dynamic_auth.keyfile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.712964] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.713136] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vendordata_dynamic_auth.timeout = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.713315] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vmware.api_retry_count = 10 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.713477] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vmware.ca_file = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.713649] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vmware.cache_prefix = devstack-image-cache {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.713816] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vmware.cluster_name = testcl1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.713982] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vmware.connection_pool_size = 10 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.714182] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vmware.console_delay_seconds = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.714378] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vmware.datastore_regex = ^datastore.* {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.714591] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.714766] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vmware.host_password = **** {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.714935] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vmware.host_port = 443 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.715119] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vmware.host_username = administrator@vsphere.local {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.715293] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vmware.insecure = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.715513] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vmware.integration_bridge = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.715655] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vmware.maximum_objects = 100 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.715817] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vmware.pbm_default_policy = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.715976] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vmware.pbm_enabled = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.716146] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vmware.pbm_wsdl_location = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.716317] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.716503] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vmware.serial_port_proxy_uri = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.716668] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vmware.serial_port_service_uri = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.716837] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vmware.task_poll_interval = 0.5 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.717017] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vmware.use_linked_clone = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.717191] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vmware.vnc_keymap = en-us {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.717359] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vmware.vnc_port = 5900 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.717526] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vmware.vnc_port_total = 10000 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.717715] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vnc.auth_schemes = ['none'] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.717892] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vnc.enabled = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.718209] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.718400] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.718571] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vnc.novncproxy_port = 6080 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.718749] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vnc.server_listen = 127.0.0.1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.718923] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.719096] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vnc.vencrypt_ca_certs = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.719261] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vnc.vencrypt_client_cert = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.719421] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vnc.vencrypt_client_key = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.719603] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.719769] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] workarounds.disable_deep_image_inspection = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.719934] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.720109] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.720275] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.720453] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] workarounds.disable_rootwrap = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.720649] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] workarounds.enable_numa_live_migration = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.720822] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.720986] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.721163] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.721330] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] workarounds.libvirt_disable_apic = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.721494] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.721662] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.721824] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.721989] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.722166] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.722328] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.722491] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.722655] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.722817] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.722983] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.723185] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.723359] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] wsgi.client_socket_timeout = 900 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.723530] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] wsgi.default_pool_size = 1000 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.723698] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] wsgi.keep_alive = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.723863] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] wsgi.max_header_line = 16384 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.724038] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] wsgi.secure_proxy_ssl_header = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.724208] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] wsgi.ssl_ca_file = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.724371] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] wsgi.ssl_cert_file = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.724576] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] wsgi.ssl_key_file = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.724700] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] wsgi.tcp_keepidle = 600 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.724875] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.725051] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] zvm.ca_file = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.725219] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] zvm.cloud_connector_url = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.725535] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.725722] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] zvm.reachable_timeout = 300 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.725909] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_policy.enforce_new_defaults = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.726310] env[62692]: WARNING oslo_config.cfg [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 559.726585] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_policy.enforce_scope = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.726802] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_policy.policy_default_rule = default {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.726998] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.727195] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_policy.policy_file = policy.yaml {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.727376] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.727541] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.727706] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.727867] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.728042] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.728217] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.728397] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.728586] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] profiler.connection_string = messaging:// {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.728767] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] profiler.enabled = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.728941] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] profiler.es_doc_type = notification {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.729123] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] profiler.es_scroll_size = 10000 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.729298] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] profiler.es_scroll_time = 2m {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.729464] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] profiler.filter_error_trace = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.729636] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] profiler.hmac_keys = **** {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.729806] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] profiler.sentinel_service_name = mymaster {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.729977] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] profiler.socket_timeout = 0.1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.730166] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] profiler.trace_requests = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.730335] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] profiler.trace_sqlalchemy = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.730517] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] profiler_jaeger.process_tags = {} {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.730681] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] profiler_jaeger.service_name_prefix = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.730845] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] profiler_otlp.service_name_prefix = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.731024] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] remote_debug.host = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.731187] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] remote_debug.port = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.731370] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.731536] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.731702] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.731860] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.732033] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.732200] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.732362] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.732530] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.732692] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.732861] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.733029] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.733206] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.733376] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.733548] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.733718] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.733884] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.734059] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.734239] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.734405] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.734572] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.734740] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.734903] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.735077] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.735247] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.735431] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.735606] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.735775] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.735935] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.736120] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.736291] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.ssl = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.736491] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.736673] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.736841] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.737014] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.737193] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.ssl_version = {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.737360] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.737555] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.737727] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_notifications.retry = -1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.737913] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.738101] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_messaging_notifications.transport_url = **** {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.738279] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_limit.auth_section = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.738444] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_limit.auth_type = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.738618] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_limit.cafile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.738808] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_limit.certfile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.738980] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_limit.collect_timing = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.739157] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_limit.connect_retries = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.739319] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_limit.connect_retry_delay = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.739479] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_limit.endpoint_id = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.739640] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_limit.endpoint_override = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.739803] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_limit.insecure = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.739961] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_limit.keyfile = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.740133] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_limit.max_version = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.740293] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_limit.min_version = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.740450] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_limit.region_name = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.740613] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_limit.retriable_status_codes = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.740770] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_limit.service_name = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.740926] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_limit.service_type = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.741100] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_limit.split_loggers = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.741263] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_limit.status_code_retries = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.741424] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_limit.status_code_retry_delay = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.741583] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_limit.timeout = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.741745] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_limit.valid_interfaces = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.741903] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_limit.version = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.742085] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_reports.file_event_handler = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.742256] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.742417] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] oslo_reports.log_dir = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.742594] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.742755] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.742917] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.743160] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.743349] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.743515] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.743693] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.743856] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vif_plug_ovs_privileged.group = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.744027] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.744203] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.744370] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.744534] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] vif_plug_ovs_privileged.user = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.744710] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] os_vif_linux_bridge.flat_interface = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.744963] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.745172] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.745355] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.745555] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.745733] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.745907] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.746085] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.746273] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.746488] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] os_vif_ovs.isolate_vif = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.746695] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.746875] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.747061] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.747240] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] os_vif_ovs.ovsdb_interface = native {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.747407] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] os_vif_ovs.per_port_bridge = False {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.747576] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] os_brick.lock_path = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.747747] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.747911] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.748098] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] privsep_osbrick.capabilities = [21] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.748265] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] privsep_osbrick.group = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.748426] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] privsep_osbrick.helper_command = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.748593] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.748759] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.748918] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] privsep_osbrick.user = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.749109] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.749274] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] nova_sys_admin.group = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.749435] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] nova_sys_admin.helper_command = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.749669] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.749865] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.750044] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] nova_sys_admin.user = None {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 559.750185] env[62692]: DEBUG oslo_service.service [None req-051f2f6b-c4d5-453c-853f-650d8d84b1a0 None None] ******************************************************************************** {{(pid=62692) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 559.750704] env[62692]: INFO nova.service [-] Starting compute node (version 0.1.0) [ 560.256859] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] Getting list of instances from cluster (obj){ [ 560.256859] env[62692]: value = "domain-c8" [ 560.256859] env[62692]: _type = "ClusterComputeResource" [ 560.256859] env[62692]: } {{(pid=62692) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 560.258055] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bf535ce-8d63-431c-80dd-edd2a6b45c9e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.267033] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] Got total of 0 instances {{(pid=62692) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 560.267578] env[62692]: WARNING nova.virt.vmwareapi.driver [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 560.268065] env[62692]: INFO nova.virt.node [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] Generated node identity 470758a8-5efa-4875-aad5-f512a727752f [ 560.268290] env[62692]: INFO nova.virt.node [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] Wrote node identity 470758a8-5efa-4875-aad5-f512a727752f to /opt/stack/data/n-cpu-1/compute_id [ 560.771258] env[62692]: WARNING nova.compute.manager [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] Compute nodes ['470758a8-5efa-4875-aad5-f512a727752f'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 561.777395] env[62692]: INFO nova.compute.manager [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 562.786821] env[62692]: WARNING nova.compute.manager [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 562.787197] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.787355] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.787505] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 562.787668] env[62692]: DEBUG nova.compute.resource_tracker [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62692) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 562.788646] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91f92439-af98-4550-9a42-86c8f91a2a7d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.797138] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22ffd8b3-bdc7-417c-8b6b-c4c4e8f1ee07 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.810816] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f25143c-6f0e-4f48-acb5-2d281ceafb77 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.816762] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a643ff82-09b1-4137-a862-d43a13b2b685 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.845997] env[62692]: DEBUG nova.compute.resource_tracker [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181536MB free_disk=167GB free_vcpus=48 pci_devices=None {{(pid=62692) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 562.846163] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.846356] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.349063] env[62692]: WARNING nova.compute.resource_tracker [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] No compute node record for cpu-1:470758a8-5efa-4875-aad5-f512a727752f: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 470758a8-5efa-4875-aad5-f512a727752f could not be found. [ 563.852757] env[62692]: INFO nova.compute.resource_tracker [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 470758a8-5efa-4875-aad5-f512a727752f [ 565.364306] env[62692]: DEBUG nova.compute.resource_tracker [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62692) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 565.364640] env[62692]: DEBUG nova.compute.resource_tracker [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62692) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 565.529960] env[62692]: INFO nova.scheduler.client.report [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] [req-4edce56b-74db-4048-ab6b-8064fbbf2b1c] Created resource provider record via placement API for resource provider with UUID 470758a8-5efa-4875-aad5-f512a727752f and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 565.547008] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8fd4da1-f6d4-4db5-b2ba-e837056c7ef4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.554575] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b037830e-2e72-4300-a69e-6ceb993f2e8d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.584761] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-514e3034-eb2f-46b8-b95a-91c951f9a524 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.591746] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bd6f1f1-a96a-47cb-8fe7-1ebc53643884 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.604250] env[62692]: DEBUG nova.compute.provider_tree [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 566.138233] env[62692]: DEBUG nova.scheduler.client.report [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] Updated inventory for provider 470758a8-5efa-4875-aad5-f512a727752f with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 566.138462] env[62692]: DEBUG nova.compute.provider_tree [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] Updating resource provider 470758a8-5efa-4875-aad5-f512a727752f generation from 0 to 1 during operation: update_inventory {{(pid=62692) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 566.138602] env[62692]: DEBUG nova.compute.provider_tree [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 566.188987] env[62692]: DEBUG nova.compute.provider_tree [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] Updating resource provider 470758a8-5efa-4875-aad5-f512a727752f generation from 1 to 2 during operation: update_traits {{(pid=62692) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 566.693967] env[62692]: DEBUG nova.compute.resource_tracker [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62692) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 566.694357] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.848s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.694357] env[62692]: DEBUG nova.service [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] Creating RPC server for service compute {{(pid=62692) start /opt/stack/nova/nova/service.py:186}} [ 566.707939] env[62692]: DEBUG nova.service [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] Join ServiceGroup membership for this service compute {{(pid=62692) start /opt/stack/nova/nova/service.py:203}} [ 566.708141] env[62692]: DEBUG nova.servicegroup.drivers.db [None req-6b1e0f37-4b3f-452c-8390-82f1fc3a8468 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=62692) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 603.035157] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "549f672f-f21b-46b1-9002-42a634253624" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.038126] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "549f672f-f21b-46b1-9002-42a634253624" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.541654] env[62692]: DEBUG nova.compute.manager [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 603.854428] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Acquiring lock "7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.855151] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Lock "7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.079530] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.079978] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.081292] env[62692]: INFO nova.compute.claims [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 604.356135] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "a16b5c50-a48a-4dfb-b2cb-bfe3723d108e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.357144] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "a16b5c50-a48a-4dfb-b2cb-bfe3723d108e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.359435] env[62692]: DEBUG nova.compute.manager [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 604.495930] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Acquiring lock "bf07e236-28b8-416e-94d3-fdf75cf62c5b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.496805] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Lock "bf07e236-28b8-416e-94d3-fdf75cf62c5b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.858736] env[62692]: DEBUG nova.compute.manager [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 604.889050] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.999559] env[62692]: DEBUG nova.compute.manager [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 605.179018] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d48c784-7da8-4757-8e5a-c079365d0b18 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.187508] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3319a3b4-27c7-4f77-aeb1-08307f29bb95 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.222729] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b6ca97e-48a5-45a3-85c3-6d19cea658b2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.230698] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab4280f2-cb4a-41d4-a0e4-04634f9d06a2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.247215] env[62692]: DEBUG nova.compute.provider_tree [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 605.388975] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.514080] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Acquiring lock "2d5a412a-66c0-46fc-84bc-2edb90f4593c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.514340] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Lock "2d5a412a-66c0-46fc-84bc-2edb90f4593c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.546218] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.751614] env[62692]: DEBUG nova.scheduler.client.report [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 606.023469] env[62692]: DEBUG nova.compute.manager [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 606.262435] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.183s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.263138] env[62692]: DEBUG nova.compute.manager [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 606.268815] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.378s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.268815] env[62692]: INFO nova.compute.claims [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 606.556451] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.774606] env[62692]: DEBUG nova.compute.utils [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 606.780061] env[62692]: DEBUG nova.compute.manager [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 606.780061] env[62692]: DEBUG nova.network.neutron [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 606.960390] env[62692]: DEBUG nova.policy [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f2e2ee26163841ddb15bb67520e98394', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e13f8888d5e84eed8247101558e29a56', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 607.287595] env[62692]: DEBUG nova.compute.manager [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 607.404921] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04b25403-7c5e-4479-ad13-f0820d671605 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.414741] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c1ac2e4-c985-4634-ae82-2edb055f8f00 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.454312] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-757f5e83-78f5-4573-a6b8-27f723ca02df {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.462095] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-243ec54a-f5d7-4ef7-9626-1fb4b43f170c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.477953] env[62692]: DEBUG nova.compute.provider_tree [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 607.633975] env[62692]: DEBUG oslo_concurrency.lockutils [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Acquiring lock "331411c2-1758-46a4-ae08-8c575d0b8be2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.634279] env[62692]: DEBUG oslo_concurrency.lockutils [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Lock "331411c2-1758-46a4-ae08-8c575d0b8be2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.658904] env[62692]: DEBUG nova.network.neutron [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Successfully created port: fcb13275-a594-4d83-b081-729b1c23b331 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 607.983162] env[62692]: DEBUG nova.scheduler.client.report [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 608.138260] env[62692]: DEBUG nova.compute.manager [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 608.300250] env[62692]: DEBUG nova.compute.manager [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 608.336361] env[62692]: DEBUG nova.virt.hardware [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 608.337028] env[62692]: DEBUG nova.virt.hardware [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 608.337197] env[62692]: DEBUG nova.virt.hardware [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 608.337385] env[62692]: DEBUG nova.virt.hardware [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 608.337554] env[62692]: DEBUG nova.virt.hardware [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 608.337720] env[62692]: DEBUG nova.virt.hardware [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 608.337951] env[62692]: DEBUG nova.virt.hardware [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 608.338126] env[62692]: DEBUG nova.virt.hardware [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 608.338541] env[62692]: DEBUG nova.virt.hardware [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 608.338719] env[62692]: DEBUG nova.virt.hardware [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 608.338909] env[62692]: DEBUG nova.virt.hardware [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 608.339799] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6d76406-87a8-43f6-a623-4063e367dc7e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.348589] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-685180f2-bbc8-4716-9df2-9fa4f33e5831 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.370540] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3c077d7-4eaf-457d-883e-2ed0488e69eb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.489683] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.223s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.490291] env[62692]: DEBUG nova.compute.manager [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 608.493348] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.104s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.499119] env[62692]: INFO nova.compute.claims [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 608.664904] env[62692]: DEBUG oslo_concurrency.lockutils [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.710120] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._sync_power_states {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 609.003910] env[62692]: DEBUG nova.compute.utils [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 609.012673] env[62692]: DEBUG nova.compute.manager [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 609.013040] env[62692]: DEBUG nova.network.neutron [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 609.132856] env[62692]: DEBUG nova.policy [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b52c2aa90394303a127d9ece53cd78f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa15800872af4c2393abee0ec1e43d59', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 609.217388] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Getting list of instances from cluster (obj){ [ 609.217388] env[62692]: value = "domain-c8" [ 609.217388] env[62692]: _type = "ClusterComputeResource" [ 609.217388] env[62692]: } {{(pid=62692) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 609.219185] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d2088d1-52f8-420f-9431-a3a9b9c82996 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.229758] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Got total of 0 instances {{(pid=62692) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 609.230021] env[62692]: WARNING nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] While synchronizing instance power states, found 3 instances in the database and 0 instances on the hypervisor. [ 609.232096] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Triggering sync for uuid 549f672f-f21b-46b1-9002-42a634253624 {{(pid=62692) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 609.232096] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Triggering sync for uuid 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6 {{(pid=62692) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 609.232096] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Triggering sync for uuid a16b5c50-a48a-4dfb-b2cb-bfe3723d108e {{(pid=62692) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 609.232096] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "549f672f-f21b-46b1-9002-42a634253624" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.232096] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.232096] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "a16b5c50-a48a-4dfb-b2cb-bfe3723d108e" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.232320] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 609.232320] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Getting list of instances from cluster (obj){ [ 609.232320] env[62692]: value = "domain-c8" [ 609.232320] env[62692]: _type = "ClusterComputeResource" [ 609.232320] env[62692]: } {{(pid=62692) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 609.233054] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05146250-74c7-4235-aa5a-74c447936a1a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.241783] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Got total of 0 instances {{(pid=62692) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 609.521513] env[62692]: DEBUG nova.compute.manager [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 609.652355] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-354d0986-acbe-4545-a1ee-dca6baba6802 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.660566] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b27c305-fe4d-4aa2-91c9-1198755a0df0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.706934] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5070cb6a-9634-40f8-a4c6-965e71ef26ff {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.716123] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3d1fe6b-950d-4cdf-8b2b-1fd144cfb7ea {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.730797] env[62692]: DEBUG nova.compute.provider_tree [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.237685] env[62692]: DEBUG nova.scheduler.client.report [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 610.505484] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Acquiring lock "5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.505484] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Lock "5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.533402] env[62692]: DEBUG nova.compute.manager [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 610.538894] env[62692]: ERROR nova.compute.manager [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fcb13275-a594-4d83-b081-729b1c23b331, please check neutron logs for more information. [ 610.538894] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 610.538894] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.538894] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 610.538894] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 610.538894] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 610.538894] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 610.538894] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 610.538894] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.538894] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 610.538894] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.538894] env[62692]: ERROR nova.compute.manager raise self.value [ 610.538894] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 610.538894] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 610.538894] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.538894] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 610.539651] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.539651] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 610.539651] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fcb13275-a594-4d83-b081-729b1c23b331, please check neutron logs for more information. [ 610.539651] env[62692]: ERROR nova.compute.manager [ 610.539651] env[62692]: Traceback (most recent call last): [ 610.539651] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 610.539651] env[62692]: listener.cb(fileno) [ 610.539651] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.539651] env[62692]: result = function(*args, **kwargs) [ 610.539651] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 610.539651] env[62692]: return func(*args, **kwargs) [ 610.539651] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.539651] env[62692]: raise e [ 610.539651] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.539651] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 610.539651] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 610.539651] env[62692]: created_port_ids = self._update_ports_for_instance( [ 610.539651] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 610.539651] env[62692]: with excutils.save_and_reraise_exception(): [ 610.539651] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.539651] env[62692]: self.force_reraise() [ 610.539651] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.539651] env[62692]: raise self.value [ 610.539651] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 610.539651] env[62692]: updated_port = self._update_port( [ 610.539651] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.539651] env[62692]: _ensure_no_port_binding_failure(port) [ 610.539651] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.539651] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 610.541379] env[62692]: nova.exception.PortBindingFailed: Binding failed for port fcb13275-a594-4d83-b081-729b1c23b331, please check neutron logs for more information. [ 610.541379] env[62692]: Removing descriptor: 15 [ 610.541379] env[62692]: ERROR nova.compute.manager [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fcb13275-a594-4d83-b081-729b1c23b331, please check neutron logs for more information. [ 610.541379] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] Traceback (most recent call last): [ 610.541379] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 610.541379] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] yield resources [ 610.541379] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 610.541379] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] self.driver.spawn(context, instance, image_meta, [ 610.541379] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 610.541379] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] self._vmops.spawn(context, instance, image_meta, injected_files, [ 610.541379] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 610.541379] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] vm_ref = self.build_virtual_machine(instance, [ 610.541705] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 610.541705] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] vif_infos = vmwarevif.get_vif_info(self._session, [ 610.541705] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 610.541705] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] for vif in network_info: [ 610.541705] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 610.541705] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] return self._sync_wrapper(fn, *args, **kwargs) [ 610.541705] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 610.541705] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] self.wait() [ 610.541705] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 610.541705] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] self[:] = self._gt.wait() [ 610.541705] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 610.541705] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] return self._exit_event.wait() [ 610.541705] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 610.542058] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] result = hub.switch() [ 610.542058] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 610.542058] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] return self.greenlet.switch() [ 610.542058] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.542058] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] result = function(*args, **kwargs) [ 610.542058] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 610.542058] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] return func(*args, **kwargs) [ 610.542058] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.542058] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] raise e [ 610.542058] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.542058] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] nwinfo = self.network_api.allocate_for_instance( [ 610.542058] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 610.542058] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] created_port_ids = self._update_ports_for_instance( [ 610.542792] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 610.542792] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] with excutils.save_and_reraise_exception(): [ 610.542792] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.542792] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] self.force_reraise() [ 610.542792] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.542792] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] raise self.value [ 610.542792] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 610.542792] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] updated_port = self._update_port( [ 610.542792] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.542792] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] _ensure_no_port_binding_failure(port) [ 610.542792] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.542792] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] raise exception.PortBindingFailed(port_id=port['id']) [ 610.543120] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] nova.exception.PortBindingFailed: Binding failed for port fcb13275-a594-4d83-b081-729b1c23b331, please check neutron logs for more information. [ 610.543120] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] [ 610.543120] env[62692]: INFO nova.compute.manager [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Terminating instance [ 610.543120] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "refresh_cache-549f672f-f21b-46b1-9002-42a634253624" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.543120] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquired lock "refresh_cache-549f672f-f21b-46b1-9002-42a634253624" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.543265] env[62692]: DEBUG nova.network.neutron [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 610.565177] env[62692]: DEBUG nova.virt.hardware [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 610.565462] env[62692]: DEBUG nova.virt.hardware [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 610.565619] env[62692]: DEBUG nova.virt.hardware [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 610.565817] env[62692]: DEBUG nova.virt.hardware [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 610.566012] env[62692]: DEBUG nova.virt.hardware [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 610.566644] env[62692]: DEBUG nova.virt.hardware [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 610.566817] env[62692]: DEBUG nova.virt.hardware [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 610.566949] env[62692]: DEBUG nova.virt.hardware [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 610.567141] env[62692]: DEBUG nova.virt.hardware [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 610.567307] env[62692]: DEBUG nova.virt.hardware [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 610.567494] env[62692]: DEBUG nova.virt.hardware [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 610.568704] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d9b3995-d861-4057-95cb-b990cca32b28 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.578951] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63ed4ad2-584a-41b9-bd8c-7164381a4914 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.743593] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.250s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.744322] env[62692]: DEBUG nova.compute.manager [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 610.751225] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.203s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.751225] env[62692]: INFO nova.compute.claims [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 610.844960] env[62692]: DEBUG nova.network.neutron [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Successfully created port: 4312822f-3bf3-49ed-864c-fd84521b74ee {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 611.013593] env[62692]: DEBUG nova.compute.manager [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 611.098480] env[62692]: DEBUG nova.network.neutron [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 611.262146] env[62692]: DEBUG nova.compute.utils [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 611.263182] env[62692]: DEBUG nova.compute.manager [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 611.263796] env[62692]: DEBUG nova.network.neutron [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 611.309019] env[62692]: DEBUG nova.policy [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '249a3b566fcf4d88a973693512df370e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0b0b2e09a06a4fc7a00a03f72989ffed', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 611.320870] env[62692]: DEBUG nova.network.neutron [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.551368] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.747277] env[62692]: DEBUG nova.network.neutron [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Successfully created port: 8a34d418-e31d-40da-b68f-889fef817fb8 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 611.770381] env[62692]: DEBUG nova.compute.manager [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 611.825344] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Releasing lock "refresh_cache-549f672f-f21b-46b1-9002-42a634253624" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.825791] env[62692]: DEBUG nova.compute.manager [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 611.826353] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 611.827210] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b9865a84-ccb2-4294-92e4-7fe848f555b9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.836361] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baa2e419-12fa-45fc-a2a0-b0af0988ce3f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.866158] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 549f672f-f21b-46b1-9002-42a634253624 could not be found. [ 611.866158] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 611.866158] env[62692]: INFO nova.compute.manager [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Took 0.04 seconds to destroy the instance on the hypervisor. [ 611.866158] env[62692]: DEBUG oslo.service.loopingcall [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 611.866158] env[62692]: DEBUG nova.compute.manager [-] [instance: 549f672f-f21b-46b1-9002-42a634253624] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 611.866158] env[62692]: DEBUG nova.network.neutron [-] [instance: 549f672f-f21b-46b1-9002-42a634253624] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 611.906680] env[62692]: DEBUG nova.network.neutron [-] [instance: 549f672f-f21b-46b1-9002-42a634253624] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 611.917483] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68b201f9-8773-4abd-9bd7-b64c80e2c87e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.925197] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-231c0a66-3c1a-4a02-8bf4-7cc5030ef026 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.959486] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c4149f1-b65a-481c-9433-35ca6d536feb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.966954] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20046c9f-0520-45d9-a93d-d24ad9e4bf99 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.980811] env[62692]: DEBUG nova.compute.provider_tree [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 612.360582] env[62692]: DEBUG nova.compute.manager [req-4d0ed439-fc9c-4af2-b6dd-12d1ba306196 req-e31d60b9-f365-40ce-af81-8489a1d8a88a service nova] [instance: 549f672f-f21b-46b1-9002-42a634253624] Received event network-changed-fcb13275-a594-4d83-b081-729b1c23b331 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 612.360784] env[62692]: DEBUG nova.compute.manager [req-4d0ed439-fc9c-4af2-b6dd-12d1ba306196 req-e31d60b9-f365-40ce-af81-8489a1d8a88a service nova] [instance: 549f672f-f21b-46b1-9002-42a634253624] Refreshing instance network info cache due to event network-changed-fcb13275-a594-4d83-b081-729b1c23b331. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 612.361014] env[62692]: DEBUG oslo_concurrency.lockutils [req-4d0ed439-fc9c-4af2-b6dd-12d1ba306196 req-e31d60b9-f365-40ce-af81-8489a1d8a88a service nova] Acquiring lock "refresh_cache-549f672f-f21b-46b1-9002-42a634253624" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.361525] env[62692]: DEBUG oslo_concurrency.lockutils [req-4d0ed439-fc9c-4af2-b6dd-12d1ba306196 req-e31d60b9-f365-40ce-af81-8489a1d8a88a service nova] Acquired lock "refresh_cache-549f672f-f21b-46b1-9002-42a634253624" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.361525] env[62692]: DEBUG nova.network.neutron [req-4d0ed439-fc9c-4af2-b6dd-12d1ba306196 req-e31d60b9-f365-40ce-af81-8489a1d8a88a service nova] [instance: 549f672f-f21b-46b1-9002-42a634253624] Refreshing network info cache for port fcb13275-a594-4d83-b081-729b1c23b331 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 612.410247] env[62692]: DEBUG nova.network.neutron [-] [instance: 549f672f-f21b-46b1-9002-42a634253624] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.485957] env[62692]: DEBUG nova.scheduler.client.report [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 612.781859] env[62692]: DEBUG nova.compute.manager [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 612.810223] env[62692]: DEBUG nova.virt.hardware [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 612.810485] env[62692]: DEBUG nova.virt.hardware [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 612.810640] env[62692]: DEBUG nova.virt.hardware [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 612.810825] env[62692]: DEBUG nova.virt.hardware [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 612.810974] env[62692]: DEBUG nova.virt.hardware [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 612.811474] env[62692]: DEBUG nova.virt.hardware [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 612.811689] env[62692]: DEBUG nova.virt.hardware [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 612.811798] env[62692]: DEBUG nova.virt.hardware [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 612.812062] env[62692]: DEBUG nova.virt.hardware [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 612.812690] env[62692]: DEBUG nova.virt.hardware [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 612.812939] env[62692]: DEBUG nova.virt.hardware [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 612.813824] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb6629ec-4b2e-4b71-a397-f474fcf7385a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.822014] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-521fe72e-3275-4563-acbd-4e90fc707264 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.829060] env[62692]: ERROR nova.compute.manager [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4312822f-3bf3-49ed-864c-fd84521b74ee, please check neutron logs for more information. [ 612.829060] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 612.829060] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 612.829060] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 612.829060] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 612.829060] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 612.829060] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 612.829060] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 612.829060] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 612.829060] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 612.829060] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 612.829060] env[62692]: ERROR nova.compute.manager raise self.value [ 612.829060] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 612.829060] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 612.829060] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 612.829060] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 612.829542] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 612.829542] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 612.829542] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4312822f-3bf3-49ed-864c-fd84521b74ee, please check neutron logs for more information. [ 612.829542] env[62692]: ERROR nova.compute.manager [ 612.829542] env[62692]: Traceback (most recent call last): [ 612.829542] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 612.829542] env[62692]: listener.cb(fileno) [ 612.829542] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 612.829542] env[62692]: result = function(*args, **kwargs) [ 612.829542] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 612.829542] env[62692]: return func(*args, **kwargs) [ 612.829542] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 612.829542] env[62692]: raise e [ 612.829542] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 612.829542] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 612.829542] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 612.829542] env[62692]: created_port_ids = self._update_ports_for_instance( [ 612.829542] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 612.829542] env[62692]: with excutils.save_and_reraise_exception(): [ 612.829542] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 612.829542] env[62692]: self.force_reraise() [ 612.829542] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 612.829542] env[62692]: raise self.value [ 612.829542] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 612.829542] env[62692]: updated_port = self._update_port( [ 612.829542] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 612.829542] env[62692]: _ensure_no_port_binding_failure(port) [ 612.829542] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 612.829542] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 612.830347] env[62692]: nova.exception.PortBindingFailed: Binding failed for port 4312822f-3bf3-49ed-864c-fd84521b74ee, please check neutron logs for more information. [ 612.830347] env[62692]: Removing descriptor: 16 [ 612.830347] env[62692]: ERROR nova.compute.manager [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4312822f-3bf3-49ed-864c-fd84521b74ee, please check neutron logs for more information. [ 612.830347] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Traceback (most recent call last): [ 612.830347] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 612.830347] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] yield resources [ 612.830347] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 612.830347] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] self.driver.spawn(context, instance, image_meta, [ 612.830347] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 612.830347] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 612.830347] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 612.830347] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] vm_ref = self.build_virtual_machine(instance, [ 612.830680] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 612.830680] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] vif_infos = vmwarevif.get_vif_info(self._session, [ 612.830680] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 612.830680] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] for vif in network_info: [ 612.830680] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 612.830680] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] return self._sync_wrapper(fn, *args, **kwargs) [ 612.830680] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 612.830680] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] self.wait() [ 612.830680] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 612.830680] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] self[:] = self._gt.wait() [ 612.830680] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 612.830680] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] return self._exit_event.wait() [ 612.830680] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 612.831063] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] result = hub.switch() [ 612.831063] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 612.831063] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] return self.greenlet.switch() [ 612.831063] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 612.831063] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] result = function(*args, **kwargs) [ 612.831063] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 612.831063] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] return func(*args, **kwargs) [ 612.831063] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 612.831063] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] raise e [ 612.831063] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 612.831063] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] nwinfo = self.network_api.allocate_for_instance( [ 612.831063] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 612.831063] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] created_port_ids = self._update_ports_for_instance( [ 612.831595] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 612.831595] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] with excutils.save_and_reraise_exception(): [ 612.831595] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 612.831595] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] self.force_reraise() [ 612.831595] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 612.831595] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] raise self.value [ 612.831595] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 612.831595] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] updated_port = self._update_port( [ 612.831595] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 612.831595] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] _ensure_no_port_binding_failure(port) [ 612.831595] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 612.831595] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] raise exception.PortBindingFailed(port_id=port['id']) [ 612.832378] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] nova.exception.PortBindingFailed: Binding failed for port 4312822f-3bf3-49ed-864c-fd84521b74ee, please check neutron logs for more information. [ 612.832378] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] [ 612.832378] env[62692]: INFO nova.compute.manager [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Terminating instance [ 612.833503] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Acquiring lock "refresh_cache-7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.833503] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Acquired lock "refresh_cache-7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.833503] env[62692]: DEBUG nova.network.neutron [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 612.882887] env[62692]: DEBUG nova.network.neutron [req-4d0ed439-fc9c-4af2-b6dd-12d1ba306196 req-e31d60b9-f365-40ce-af81-8489a1d8a88a service nova] [instance: 549f672f-f21b-46b1-9002-42a634253624] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 612.914642] env[62692]: INFO nova.compute.manager [-] [instance: 549f672f-f21b-46b1-9002-42a634253624] Took 1.05 seconds to deallocate network for instance. [ 612.919334] env[62692]: DEBUG nova.compute.claims [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 612.919744] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.996044] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.243s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.996044] env[62692]: DEBUG nova.compute.manager [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 612.997095] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.441s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.998405] env[62692]: INFO nova.compute.claims [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 613.134395] env[62692]: ERROR nova.compute.manager [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8a34d418-e31d-40da-b68f-889fef817fb8, please check neutron logs for more information. [ 613.134395] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 613.134395] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.134395] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 613.134395] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 613.134395] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 613.134395] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 613.134395] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 613.134395] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.134395] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 613.134395] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.134395] env[62692]: ERROR nova.compute.manager raise self.value [ 613.134395] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 613.134395] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 613.134395] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.134395] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 613.134989] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.134989] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 613.134989] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8a34d418-e31d-40da-b68f-889fef817fb8, please check neutron logs for more information. [ 613.134989] env[62692]: ERROR nova.compute.manager [ 613.134989] env[62692]: Traceback (most recent call last): [ 613.134989] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 613.134989] env[62692]: listener.cb(fileno) [ 613.134989] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.134989] env[62692]: result = function(*args, **kwargs) [ 613.134989] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 613.134989] env[62692]: return func(*args, **kwargs) [ 613.134989] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 613.134989] env[62692]: raise e [ 613.134989] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.134989] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 613.134989] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 613.134989] env[62692]: created_port_ids = self._update_ports_for_instance( [ 613.134989] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 613.134989] env[62692]: with excutils.save_and_reraise_exception(): [ 613.134989] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.134989] env[62692]: self.force_reraise() [ 613.134989] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.134989] env[62692]: raise self.value [ 613.134989] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 613.134989] env[62692]: updated_port = self._update_port( [ 613.134989] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.134989] env[62692]: _ensure_no_port_binding_failure(port) [ 613.134989] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.134989] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 613.135839] env[62692]: nova.exception.PortBindingFailed: Binding failed for port 8a34d418-e31d-40da-b68f-889fef817fb8, please check neutron logs for more information. [ 613.135839] env[62692]: Removing descriptor: 17 [ 613.135839] env[62692]: ERROR nova.compute.manager [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8a34d418-e31d-40da-b68f-889fef817fb8, please check neutron logs for more information. [ 613.135839] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Traceback (most recent call last): [ 613.135839] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 613.135839] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] yield resources [ 613.135839] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 613.135839] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] self.driver.spawn(context, instance, image_meta, [ 613.135839] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 613.135839] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 613.135839] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 613.135839] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] vm_ref = self.build_virtual_machine(instance, [ 613.136336] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 613.136336] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] vif_infos = vmwarevif.get_vif_info(self._session, [ 613.136336] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 613.136336] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] for vif in network_info: [ 613.136336] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 613.136336] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] return self._sync_wrapper(fn, *args, **kwargs) [ 613.136336] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 613.136336] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] self.wait() [ 613.136336] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 613.136336] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] self[:] = self._gt.wait() [ 613.136336] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 613.136336] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] return self._exit_event.wait() [ 613.136336] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 613.136797] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] result = hub.switch() [ 613.136797] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 613.136797] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] return self.greenlet.switch() [ 613.136797] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.136797] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] result = function(*args, **kwargs) [ 613.136797] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 613.136797] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] return func(*args, **kwargs) [ 613.136797] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 613.136797] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] raise e [ 613.136797] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.136797] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] nwinfo = self.network_api.allocate_for_instance( [ 613.136797] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 613.136797] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] created_port_ids = self._update_ports_for_instance( [ 613.137192] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 613.137192] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] with excutils.save_and_reraise_exception(): [ 613.137192] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.137192] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] self.force_reraise() [ 613.137192] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.137192] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] raise self.value [ 613.137192] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 613.137192] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] updated_port = self._update_port( [ 613.137192] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.137192] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] _ensure_no_port_binding_failure(port) [ 613.137192] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.137192] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] raise exception.PortBindingFailed(port_id=port['id']) [ 613.137560] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] nova.exception.PortBindingFailed: Binding failed for port 8a34d418-e31d-40da-b68f-889fef817fb8, please check neutron logs for more information. [ 613.137560] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] [ 613.137560] env[62692]: INFO nova.compute.manager [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Terminating instance [ 613.141042] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "refresh_cache-a16b5c50-a48a-4dfb-b2cb-bfe3723d108e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.141042] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquired lock "refresh_cache-a16b5c50-a48a-4dfb-b2cb-bfe3723d108e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.141042] env[62692]: DEBUG nova.network.neutron [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 613.141042] env[62692]: DEBUG nova.network.neutron [req-4d0ed439-fc9c-4af2-b6dd-12d1ba306196 req-e31d60b9-f365-40ce-af81-8489a1d8a88a service nova] [instance: 549f672f-f21b-46b1-9002-42a634253624] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.374222] env[62692]: DEBUG nova.network.neutron [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 613.484773] env[62692]: DEBUG nova.network.neutron [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.506068] env[62692]: DEBUG nova.compute.utils [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 613.515623] env[62692]: DEBUG nova.compute.manager [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 613.515623] env[62692]: DEBUG nova.network.neutron [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 613.648094] env[62692]: DEBUG oslo_concurrency.lockutils [req-4d0ed439-fc9c-4af2-b6dd-12d1ba306196 req-e31d60b9-f365-40ce-af81-8489a1d8a88a service nova] Releasing lock "refresh_cache-549f672f-f21b-46b1-9002-42a634253624" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.682984] env[62692]: DEBUG nova.network.neutron [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 613.691716] env[62692]: DEBUG nova.policy [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '110ea776273146c9b105fd9dcd7ed50e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c533116e7f04465490b10a7b66cf74ab', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 613.771512] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Acquiring lock "647d9514-5c31-4970-82e7-626e08f00237" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.771738] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Lock "647d9514-5c31-4970-82e7-626e08f00237" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.823854] env[62692]: DEBUG nova.network.neutron [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.991544] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Releasing lock "refresh_cache-7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.991544] env[62692]: DEBUG nova.compute.manager [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 613.991544] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 613.991544] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1ef7889d-01c6-45b4-8581-64cfaf94a8cc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.000334] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb04b0ce-7965-4860-a2ea-e40a7dce3a9a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.017183] env[62692]: DEBUG nova.compute.manager [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 614.034132] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Acquiring lock "4fabba8d-25f5-48a0-a844-bdec6ea5c422" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.034345] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Lock "4fabba8d-25f5-48a0-a844-bdec6ea5c422" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.044810] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6 could not be found. [ 614.044810] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 614.044810] env[62692]: INFO nova.compute.manager [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Took 0.05 seconds to destroy the instance on the hypervisor. [ 614.044810] env[62692]: DEBUG oslo.service.loopingcall [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 614.044810] env[62692]: DEBUG nova.compute.manager [-] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 614.044810] env[62692]: DEBUG nova.network.neutron [-] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 614.077202] env[62692]: DEBUG nova.network.neutron [-] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 614.209606] env[62692]: DEBUG nova.compute.manager [req-d5dd7c5a-69ea-422b-944a-2bc0ccd9bbb1 req-d34345ad-c5b3-456a-bdd8-d1c3735474bc service nova] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Received event network-changed-8a34d418-e31d-40da-b68f-889fef817fb8 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 614.209742] env[62692]: DEBUG nova.compute.manager [req-d5dd7c5a-69ea-422b-944a-2bc0ccd9bbb1 req-d34345ad-c5b3-456a-bdd8-d1c3735474bc service nova] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Refreshing instance network info cache due to event network-changed-8a34d418-e31d-40da-b68f-889fef817fb8. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 614.209936] env[62692]: DEBUG oslo_concurrency.lockutils [req-d5dd7c5a-69ea-422b-944a-2bc0ccd9bbb1 req-d34345ad-c5b3-456a-bdd8-d1c3735474bc service nova] Acquiring lock "refresh_cache-a16b5c50-a48a-4dfb-b2cb-bfe3723d108e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.229905] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57ff4333-9adf-49d9-bca4-015657d541bc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.242295] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80c3a3d9-e1b0-43a7-8a64-388719eec921 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.286407] env[62692]: DEBUG nova.compute.manager [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 614.290045] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca65b5ec-e021-44a7-86fb-8891c2dd0270 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.298105] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9972587e-868a-4a05-87b8-a8f34789f37a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.315166] env[62692]: DEBUG nova.compute.provider_tree [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 614.328939] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Releasing lock "refresh_cache-a16b5c50-a48a-4dfb-b2cb-bfe3723d108e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.328939] env[62692]: DEBUG nova.compute.manager [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 614.328939] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 614.330712] env[62692]: DEBUG oslo_concurrency.lockutils [req-d5dd7c5a-69ea-422b-944a-2bc0ccd9bbb1 req-d34345ad-c5b3-456a-bdd8-d1c3735474bc service nova] Acquired lock "refresh_cache-a16b5c50-a48a-4dfb-b2cb-bfe3723d108e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.330712] env[62692]: DEBUG nova.network.neutron [req-d5dd7c5a-69ea-422b-944a-2bc0ccd9bbb1 req-d34345ad-c5b3-456a-bdd8-d1c3735474bc service nova] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Refreshing network info cache for port 8a34d418-e31d-40da-b68f-889fef817fb8 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 614.330712] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7ec9e76c-c183-4fa9-8adc-a3d5dedd3b49 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.339735] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f36160c-c794-42ae-89d1-5209c4b1cf11 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.366780] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a16b5c50-a48a-4dfb-b2cb-bfe3723d108e could not be found. [ 614.367099] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 614.367365] env[62692]: INFO nova.compute.manager [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 614.367578] env[62692]: DEBUG oslo.service.loopingcall [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 614.367866] env[62692]: DEBUG nova.compute.manager [-] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 614.367969] env[62692]: DEBUG nova.network.neutron [-] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 614.392859] env[62692]: DEBUG nova.network.neutron [-] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 614.406923] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Acquiring lock "54f615e3-5c6e-44bb-aeb7-5620a639b55c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.407177] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Lock "54f615e3-5c6e-44bb-aeb7-5620a639b55c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.545514] env[62692]: DEBUG nova.compute.manager [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 614.553203] env[62692]: DEBUG nova.network.neutron [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Successfully created port: b7e29e04-c6d2-4a03-bfdc-f406b818af64 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 614.580044] env[62692]: DEBUG nova.network.neutron [-] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.819270] env[62692]: DEBUG nova.scheduler.client.report [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 614.823210] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.884301] env[62692]: DEBUG nova.network.neutron [req-d5dd7c5a-69ea-422b-944a-2bc0ccd9bbb1 req-d34345ad-c5b3-456a-bdd8-d1c3735474bc service nova] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 614.896248] env[62692]: DEBUG nova.network.neutron [-] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.911654] env[62692]: DEBUG nova.compute.manager [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 615.036274] env[62692]: DEBUG nova.compute.manager [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 615.069077] env[62692]: DEBUG nova.virt.hardware [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 615.069361] env[62692]: DEBUG nova.virt.hardware [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 615.069542] env[62692]: DEBUG nova.virt.hardware [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 615.069699] env[62692]: DEBUG nova.virt.hardware [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 615.069959] env[62692]: DEBUG nova.virt.hardware [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 615.070078] env[62692]: DEBUG nova.virt.hardware [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 615.070205] env[62692]: DEBUG nova.virt.hardware [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 615.070406] env[62692]: DEBUG nova.virt.hardware [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 615.071028] env[62692]: DEBUG nova.virt.hardware [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 615.071028] env[62692]: DEBUG nova.virt.hardware [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 615.071028] env[62692]: DEBUG nova.virt.hardware [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 615.072620] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-644fbb00-37dc-41a8-beed-0635fabd8c47 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.078927] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.085161] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c573777-0abd-4936-b75e-07907853ae5c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.089175] env[62692]: INFO nova.compute.manager [-] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Took 1.05 seconds to deallocate network for instance. [ 615.091764] env[62692]: DEBUG nova.compute.claims [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 615.091764] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.102058] env[62692]: DEBUG nova.network.neutron [req-d5dd7c5a-69ea-422b-944a-2bc0ccd9bbb1 req-d34345ad-c5b3-456a-bdd8-d1c3735474bc service nova] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.200450] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 615.200710] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 615.201211] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Starting heal instance info cache {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 615.201375] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Rebuilding the list of instances to heal {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 615.329191] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.331s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.329404] env[62692]: DEBUG nova.compute.manager [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 615.339157] env[62692]: DEBUG oslo_concurrency.lockutils [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.673s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.339416] env[62692]: INFO nova.compute.claims [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 615.397761] env[62692]: INFO nova.compute.manager [-] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Took 1.03 seconds to deallocate network for instance. [ 615.400734] env[62692]: DEBUG nova.compute.claims [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 615.400864] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.461459] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.548318] env[62692]: DEBUG nova.compute.manager [req-7c478661-c3da-4862-b81a-8281f00ea39d req-64470af9-a5f5-4f4f-a786-afa98f7fe2fd service nova] [instance: 549f672f-f21b-46b1-9002-42a634253624] Received event network-vif-deleted-fcb13275-a594-4d83-b081-729b1c23b331 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 615.548318] env[62692]: DEBUG nova.compute.manager [req-7c478661-c3da-4862-b81a-8281f00ea39d req-64470af9-a5f5-4f4f-a786-afa98f7fe2fd service nova] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Received event network-changed-4312822f-3bf3-49ed-864c-fd84521b74ee {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 615.548318] env[62692]: DEBUG nova.compute.manager [req-7c478661-c3da-4862-b81a-8281f00ea39d req-64470af9-a5f5-4f4f-a786-afa98f7fe2fd service nova] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Refreshing instance network info cache due to event network-changed-4312822f-3bf3-49ed-864c-fd84521b74ee. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 615.548318] env[62692]: DEBUG oslo_concurrency.lockutils [req-7c478661-c3da-4862-b81a-8281f00ea39d req-64470af9-a5f5-4f4f-a786-afa98f7fe2fd service nova] Acquiring lock "refresh_cache-7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.548318] env[62692]: DEBUG oslo_concurrency.lockutils [req-7c478661-c3da-4862-b81a-8281f00ea39d req-64470af9-a5f5-4f4f-a786-afa98f7fe2fd service nova] Acquired lock "refresh_cache-7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.548473] env[62692]: DEBUG nova.network.neutron [req-7c478661-c3da-4862-b81a-8281f00ea39d req-64470af9-a5f5-4f4f-a786-afa98f7fe2fd service nova] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Refreshing network info cache for port 4312822f-3bf3-49ed-864c-fd84521b74ee {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 615.605557] env[62692]: DEBUG oslo_concurrency.lockutils [req-d5dd7c5a-69ea-422b-944a-2bc0ccd9bbb1 req-d34345ad-c5b3-456a-bdd8-d1c3735474bc service nova] Releasing lock "refresh_cache-a16b5c50-a48a-4dfb-b2cb-bfe3723d108e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.708806] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 549f672f-f21b-46b1-9002-42a634253624] Skipping network cache update for instance because it is Building. {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 615.710018] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Skipping network cache update for instance because it is Building. {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 615.710018] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Skipping network cache update for instance because it is Building. {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 615.710018] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Skipping network cache update for instance because it is Building. {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 615.710018] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Skipping network cache update for instance because it is Building. {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 615.710018] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Didn't find any instances for network info cache update. {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 615.710018] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 615.710987] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 615.711179] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 615.713399] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 615.713399] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 615.713399] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 615.713399] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62692) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 615.713399] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager.update_available_resource {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 615.840025] env[62692]: DEBUG nova.compute.utils [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 615.841497] env[62692]: DEBUG nova.compute.manager [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 615.841581] env[62692]: DEBUG nova.network.neutron [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 615.973087] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Acquiring lock "f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.973348] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Lock "f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.041814] env[62692]: DEBUG nova.policy [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c0b62516cf1e4b189ff8ec0e411c407f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '69ca07b9babb464f9b2107dd7070f036', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 616.218129] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.219812] env[62692]: DEBUG nova.network.neutron [req-7c478661-c3da-4862-b81a-8281f00ea39d req-64470af9-a5f5-4f4f-a786-afa98f7fe2fd service nova] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 616.344833] env[62692]: DEBUG nova.compute.manager [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 616.381626] env[62692]: DEBUG nova.network.neutron [req-7c478661-c3da-4862-b81a-8281f00ea39d req-64470af9-a5f5-4f4f-a786-afa98f7fe2fd service nova] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.706570] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84260f2e-f080-4eb2-b6dd-5c17ac824b18 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.722976] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a39b729b-eb63-4d94-ad86-919ed4a0f686 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.756388] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f55872b6-cfcc-4110-9fb1-f2163f905cc3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.765410] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29abd2ce-35a3-4d19-94d5-59f8043dbdc6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.784324] env[62692]: DEBUG nova.compute.provider_tree [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 616.884508] env[62692]: DEBUG oslo_concurrency.lockutils [req-7c478661-c3da-4862-b81a-8281f00ea39d req-64470af9-a5f5-4f4f-a786-afa98f7fe2fd service nova] Releasing lock "refresh_cache-7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.884508] env[62692]: DEBUG nova.compute.manager [req-7c478661-c3da-4862-b81a-8281f00ea39d req-64470af9-a5f5-4f4f-a786-afa98f7fe2fd service nova] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Received event network-vif-deleted-4312822f-3bf3-49ed-864c-fd84521b74ee {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 617.232900] env[62692]: DEBUG nova.compute.manager [req-b6f3408a-54cb-4c1e-95be-42f0c31a2703 req-d297f868-0163-4716-ac1a-0c964ca8fd08 service nova] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Received event network-vif-deleted-8a34d418-e31d-40da-b68f-889fef817fb8 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 617.287470] env[62692]: DEBUG nova.scheduler.client.report [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 617.360105] env[62692]: DEBUG nova.compute.manager [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 617.393984] env[62692]: DEBUG nova.virt.hardware [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 617.394256] env[62692]: DEBUG nova.virt.hardware [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 617.394404] env[62692]: DEBUG nova.virt.hardware [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 617.394605] env[62692]: DEBUG nova.virt.hardware [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 617.394793] env[62692]: DEBUG nova.virt.hardware [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 617.394962] env[62692]: DEBUG nova.virt.hardware [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 617.395525] env[62692]: DEBUG nova.virt.hardware [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 617.395702] env[62692]: DEBUG nova.virt.hardware [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 617.395871] env[62692]: DEBUG nova.virt.hardware [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 617.396039] env[62692]: DEBUG nova.virt.hardware [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 617.396207] env[62692]: DEBUG nova.virt.hardware [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 617.397124] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a57ee94d-ec50-42c8-8fc1-f7f7291a1fa9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.406851] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fad8283b-5ecc-48af-818e-b86f1ef316a6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.513700] env[62692]: DEBUG nova.network.neutron [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Successfully created port: b344a4c1-5655-4faf-8f81-fa793eaf8eb6 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 617.798718] env[62692]: DEBUG oslo_concurrency.lockutils [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.461s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.799323] env[62692]: DEBUG nova.compute.manager [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 617.801990] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.251s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.807466] env[62692]: INFO nova.compute.claims [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 618.314827] env[62692]: DEBUG nova.compute.utils [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 618.315250] env[62692]: DEBUG nova.compute.manager [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 618.315535] env[62692]: DEBUG nova.network.neutron [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 618.651330] env[62692]: DEBUG nova.policy [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '74648b19171c4b04bd5271703ae2c7be', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '72d46d9d53f14f9b9e7c0da839363d4c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 618.735060] env[62692]: ERROR nova.compute.manager [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b7e29e04-c6d2-4a03-bfdc-f406b818af64, please check neutron logs for more information. [ 618.735060] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 618.735060] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.735060] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 618.735060] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 618.735060] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 618.735060] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 618.735060] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 618.735060] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.735060] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 618.735060] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.735060] env[62692]: ERROR nova.compute.manager raise self.value [ 618.735060] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 618.735060] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 618.735060] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.735060] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 618.735531] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.735531] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 618.735531] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b7e29e04-c6d2-4a03-bfdc-f406b818af64, please check neutron logs for more information. [ 618.735531] env[62692]: ERROR nova.compute.manager [ 618.735531] env[62692]: Traceback (most recent call last): [ 618.735531] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 618.735531] env[62692]: listener.cb(fileno) [ 618.735531] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.735531] env[62692]: result = function(*args, **kwargs) [ 618.735531] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 618.735531] env[62692]: return func(*args, **kwargs) [ 618.735531] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.735531] env[62692]: raise e [ 618.735531] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.735531] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 618.735531] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 618.735531] env[62692]: created_port_ids = self._update_ports_for_instance( [ 618.735531] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 618.735531] env[62692]: with excutils.save_and_reraise_exception(): [ 618.735531] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.735531] env[62692]: self.force_reraise() [ 618.735531] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.735531] env[62692]: raise self.value [ 618.735531] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 618.735531] env[62692]: updated_port = self._update_port( [ 618.735531] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.735531] env[62692]: _ensure_no_port_binding_failure(port) [ 618.735531] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.735531] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 618.736259] env[62692]: nova.exception.PortBindingFailed: Binding failed for port b7e29e04-c6d2-4a03-bfdc-f406b818af64, please check neutron logs for more information. [ 618.736259] env[62692]: Removing descriptor: 16 [ 618.736259] env[62692]: ERROR nova.compute.manager [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b7e29e04-c6d2-4a03-bfdc-f406b818af64, please check neutron logs for more information. [ 618.736259] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Traceback (most recent call last): [ 618.736259] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 618.736259] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] yield resources [ 618.736259] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 618.736259] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] self.driver.spawn(context, instance, image_meta, [ 618.736259] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 618.736259] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 618.736259] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 618.736259] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] vm_ref = self.build_virtual_machine(instance, [ 618.736641] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 618.736641] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] vif_infos = vmwarevif.get_vif_info(self._session, [ 618.736641] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 618.736641] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] for vif in network_info: [ 618.736641] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 618.736641] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] return self._sync_wrapper(fn, *args, **kwargs) [ 618.736641] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 618.736641] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] self.wait() [ 618.736641] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 618.736641] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] self[:] = self._gt.wait() [ 618.736641] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 618.736641] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] return self._exit_event.wait() [ 618.736641] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 618.737048] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] result = hub.switch() [ 618.737048] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 618.737048] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] return self.greenlet.switch() [ 618.737048] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.737048] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] result = function(*args, **kwargs) [ 618.737048] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 618.737048] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] return func(*args, **kwargs) [ 618.737048] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.737048] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] raise e [ 618.737048] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.737048] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] nwinfo = self.network_api.allocate_for_instance( [ 618.737048] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 618.737048] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] created_port_ids = self._update_ports_for_instance( [ 618.737415] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 618.737415] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] with excutils.save_and_reraise_exception(): [ 618.737415] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.737415] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] self.force_reraise() [ 618.737415] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.737415] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] raise self.value [ 618.737415] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 618.737415] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] updated_port = self._update_port( [ 618.737415] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.737415] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] _ensure_no_port_binding_failure(port) [ 618.737415] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.737415] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] raise exception.PortBindingFailed(port_id=port['id']) [ 618.737781] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] nova.exception.PortBindingFailed: Binding failed for port b7e29e04-c6d2-4a03-bfdc-f406b818af64, please check neutron logs for more information. [ 618.737781] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] [ 618.737781] env[62692]: INFO nova.compute.manager [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Terminating instance [ 618.742018] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Acquiring lock "refresh_cache-bf07e236-28b8-416e-94d3-fdf75cf62c5b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.742018] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Acquired lock "refresh_cache-bf07e236-28b8-416e-94d3-fdf75cf62c5b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.742018] env[62692]: DEBUG nova.network.neutron [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 618.829873] env[62692]: DEBUG nova.compute.manager [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 618.846981] env[62692]: DEBUG nova.compute.manager [req-ef2b6053-e6de-46c5-b4d1-db6d715971be req-fddbde04-83ac-4f1e-ba15-c86d02038f0f service nova] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Received event network-changed-b7e29e04-c6d2-4a03-bfdc-f406b818af64 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 618.846981] env[62692]: DEBUG nova.compute.manager [req-ef2b6053-e6de-46c5-b4d1-db6d715971be req-fddbde04-83ac-4f1e-ba15-c86d02038f0f service nova] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Refreshing instance network info cache due to event network-changed-b7e29e04-c6d2-4a03-bfdc-f406b818af64. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 618.846981] env[62692]: DEBUG oslo_concurrency.lockutils [req-ef2b6053-e6de-46c5-b4d1-db6d715971be req-fddbde04-83ac-4f1e-ba15-c86d02038f0f service nova] Acquiring lock "refresh_cache-bf07e236-28b8-416e-94d3-fdf75cf62c5b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.058533] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65d077c8-a327-45f5-9c6d-c9a695b401aa {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.068574] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de887f87-3a58-4977-8c28-25fae65101cb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.103616] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f321873-4b85-4950-953f-55cddf214d58 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.115025] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87094aa0-4bfd-40f3-93fa-3418fffec764 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.128307] env[62692]: DEBUG nova.compute.provider_tree [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 619.295888] env[62692]: DEBUG nova.network.neutron [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 619.561533] env[62692]: DEBUG nova.network.neutron [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.631157] env[62692]: DEBUG nova.scheduler.client.report [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 619.845177] env[62692]: DEBUG nova.compute.manager [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 619.849674] env[62692]: DEBUG nova.network.neutron [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Successfully created port: 466bf122-912c-490c-9871-bb2062bf6b3c {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 619.888784] env[62692]: DEBUG nova.virt.hardware [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 619.889077] env[62692]: DEBUG nova.virt.hardware [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 619.889281] env[62692]: DEBUG nova.virt.hardware [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 619.889474] env[62692]: DEBUG nova.virt.hardware [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 619.889624] env[62692]: DEBUG nova.virt.hardware [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 619.889766] env[62692]: DEBUG nova.virt.hardware [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 619.890612] env[62692]: DEBUG nova.virt.hardware [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 619.890612] env[62692]: DEBUG nova.virt.hardware [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 619.890612] env[62692]: DEBUG nova.virt.hardware [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 619.890745] env[62692]: DEBUG nova.virt.hardware [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 619.890830] env[62692]: DEBUG nova.virt.hardware [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 619.892336] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89b8fbd4-32f5-4173-9727-f7379ee0ab8c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.902925] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9a20295-b428-40cb-8e41-4cafb56a293f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.012937] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Acquiring lock "e6a36771-c121-4ff5-8a78-c7b7868bf95b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.013216] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Lock "e6a36771-c121-4ff5-8a78-c7b7868bf95b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.064687] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Releasing lock "refresh_cache-bf07e236-28b8-416e-94d3-fdf75cf62c5b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.065116] env[62692]: DEBUG nova.compute.manager [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 620.065316] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 620.065614] env[62692]: DEBUG oslo_concurrency.lockutils [req-ef2b6053-e6de-46c5-b4d1-db6d715971be req-fddbde04-83ac-4f1e-ba15-c86d02038f0f service nova] Acquired lock "refresh_cache-bf07e236-28b8-416e-94d3-fdf75cf62c5b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.065782] env[62692]: DEBUG nova.network.neutron [req-ef2b6053-e6de-46c5-b4d1-db6d715971be req-fddbde04-83ac-4f1e-ba15-c86d02038f0f service nova] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Refreshing network info cache for port b7e29e04-c6d2-4a03-bfdc-f406b818af64 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 620.067523] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eb60b659-c030-4c96-9278-54a3f88a9374 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.077909] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d45dd8fb-037c-4509-9840-da0c0e26130b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.102211] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bf07e236-28b8-416e-94d3-fdf75cf62c5b could not be found. [ 620.102444] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 620.102617] env[62692]: INFO nova.compute.manager [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 620.103235] env[62692]: DEBUG oslo.service.loopingcall [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 620.103235] env[62692]: DEBUG nova.compute.manager [-] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 620.103235] env[62692]: DEBUG nova.network.neutron [-] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 620.138688] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.336s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.139533] env[62692]: DEBUG nova.compute.manager [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 620.145989] env[62692]: DEBUG nova.network.neutron [-] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 620.145989] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 7.226s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.636771] env[62692]: DEBUG nova.network.neutron [req-ef2b6053-e6de-46c5-b4d1-db6d715971be req-fddbde04-83ac-4f1e-ba15-c86d02038f0f service nova] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 620.651471] env[62692]: DEBUG nova.network.neutron [-] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.653753] env[62692]: DEBUG nova.compute.utils [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 620.655554] env[62692]: DEBUG nova.compute.manager [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 620.655808] env[62692]: DEBUG nova.network.neutron [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 620.854174] env[62692]: DEBUG nova.network.neutron [req-ef2b6053-e6de-46c5-b4d1-db6d715971be req-fddbde04-83ac-4f1e-ba15-c86d02038f0f service nova] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.862273] env[62692]: DEBUG nova.policy [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8adae7ddecbc4aa28694a228f52b5bd6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b00f181776484ecebc81d7469d81153f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 620.895075] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1794ad69-35f5-453a-ae6f-d3812f4a39ad {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.904120] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b02b233c-dda8-401b-8a6d-e63d27491869 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.942325] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a75772e4-e9d5-4a5d-82c3-16d7a2a554f7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.950473] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f65d6ce7-1cfa-4753-b366-e6c8717949a4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.965837] env[62692]: DEBUG nova.compute.provider_tree [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 621.157756] env[62692]: INFO nova.compute.manager [-] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Took 1.05 seconds to deallocate network for instance. [ 621.159995] env[62692]: DEBUG nova.compute.claims [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 621.160197] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.160663] env[62692]: DEBUG nova.compute.manager [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 621.352503] env[62692]: ERROR nova.compute.manager [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b344a4c1-5655-4faf-8f81-fa793eaf8eb6, please check neutron logs for more information. [ 621.352503] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 621.352503] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.352503] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 621.352503] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 621.352503] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 621.352503] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 621.352503] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 621.352503] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.352503] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 621.352503] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.352503] env[62692]: ERROR nova.compute.manager raise self.value [ 621.352503] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 621.352503] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 621.352503] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.352503] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 621.352970] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.352970] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 621.352970] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b344a4c1-5655-4faf-8f81-fa793eaf8eb6, please check neutron logs for more information. [ 621.352970] env[62692]: ERROR nova.compute.manager [ 621.352970] env[62692]: Traceback (most recent call last): [ 621.352970] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 621.352970] env[62692]: listener.cb(fileno) [ 621.352970] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.352970] env[62692]: result = function(*args, **kwargs) [ 621.352970] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 621.352970] env[62692]: return func(*args, **kwargs) [ 621.352970] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.352970] env[62692]: raise e [ 621.352970] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.352970] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 621.352970] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 621.352970] env[62692]: created_port_ids = self._update_ports_for_instance( [ 621.352970] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 621.352970] env[62692]: with excutils.save_and_reraise_exception(): [ 621.352970] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.352970] env[62692]: self.force_reraise() [ 621.352970] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.352970] env[62692]: raise self.value [ 621.352970] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 621.352970] env[62692]: updated_port = self._update_port( [ 621.352970] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.352970] env[62692]: _ensure_no_port_binding_failure(port) [ 621.352970] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.352970] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 621.353758] env[62692]: nova.exception.PortBindingFailed: Binding failed for port b344a4c1-5655-4faf-8f81-fa793eaf8eb6, please check neutron logs for more information. [ 621.353758] env[62692]: Removing descriptor: 17 [ 621.353758] env[62692]: ERROR nova.compute.manager [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b344a4c1-5655-4faf-8f81-fa793eaf8eb6, please check neutron logs for more information. [ 621.353758] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Traceback (most recent call last): [ 621.353758] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 621.353758] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] yield resources [ 621.353758] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 621.353758] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] self.driver.spawn(context, instance, image_meta, [ 621.353758] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 621.353758] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 621.353758] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 621.353758] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] vm_ref = self.build_virtual_machine(instance, [ 621.354174] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 621.354174] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] vif_infos = vmwarevif.get_vif_info(self._session, [ 621.354174] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 621.354174] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] for vif in network_info: [ 621.354174] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 621.354174] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] return self._sync_wrapper(fn, *args, **kwargs) [ 621.354174] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 621.354174] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] self.wait() [ 621.354174] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 621.354174] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] self[:] = self._gt.wait() [ 621.354174] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 621.354174] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] return self._exit_event.wait() [ 621.354174] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 621.354533] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] result = hub.switch() [ 621.354533] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 621.354533] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] return self.greenlet.switch() [ 621.354533] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.354533] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] result = function(*args, **kwargs) [ 621.354533] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 621.354533] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] return func(*args, **kwargs) [ 621.354533] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.354533] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] raise e [ 621.354533] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.354533] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] nwinfo = self.network_api.allocate_for_instance( [ 621.354533] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 621.354533] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] created_port_ids = self._update_ports_for_instance( [ 621.354860] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 621.354860] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] with excutils.save_and_reraise_exception(): [ 621.354860] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.354860] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] self.force_reraise() [ 621.354860] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.354860] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] raise self.value [ 621.354860] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 621.354860] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] updated_port = self._update_port( [ 621.354860] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.354860] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] _ensure_no_port_binding_failure(port) [ 621.354860] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.354860] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] raise exception.PortBindingFailed(port_id=port['id']) [ 621.355284] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] nova.exception.PortBindingFailed: Binding failed for port b344a4c1-5655-4faf-8f81-fa793eaf8eb6, please check neutron logs for more information. [ 621.355284] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] [ 621.355284] env[62692]: INFO nova.compute.manager [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Terminating instance [ 621.359049] env[62692]: DEBUG oslo_concurrency.lockutils [req-ef2b6053-e6de-46c5-b4d1-db6d715971be req-fddbde04-83ac-4f1e-ba15-c86d02038f0f service nova] Releasing lock "refresh_cache-bf07e236-28b8-416e-94d3-fdf75cf62c5b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 621.361366] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Acquiring lock "refresh_cache-2d5a412a-66c0-46fc-84bc-2edb90f4593c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.362160] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Acquired lock "refresh_cache-2d5a412a-66c0-46fc-84bc-2edb90f4593c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.362862] env[62692]: DEBUG nova.network.neutron [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 621.469311] env[62692]: DEBUG nova.scheduler.client.report [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 621.943327] env[62692]: DEBUG nova.network.neutron [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 621.982919] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.837s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.983625] env[62692]: ERROR nova.compute.manager [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fcb13275-a594-4d83-b081-729b1c23b331, please check neutron logs for more information. [ 621.983625] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] Traceback (most recent call last): [ 621.983625] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 621.983625] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] self.driver.spawn(context, instance, image_meta, [ 621.983625] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 621.983625] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] self._vmops.spawn(context, instance, image_meta, injected_files, [ 621.983625] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 621.983625] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] vm_ref = self.build_virtual_machine(instance, [ 621.983625] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 621.983625] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] vif_infos = vmwarevif.get_vif_info(self._session, [ 621.983625] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 621.983990] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] for vif in network_info: [ 621.983990] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 621.983990] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] return self._sync_wrapper(fn, *args, **kwargs) [ 621.983990] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 621.983990] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] self.wait() [ 621.983990] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 621.983990] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] self[:] = self._gt.wait() [ 621.983990] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 621.983990] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] return self._exit_event.wait() [ 621.983990] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 621.983990] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] result = hub.switch() [ 621.983990] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 621.983990] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] return self.greenlet.switch() [ 621.984359] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.984359] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] result = function(*args, **kwargs) [ 621.984359] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 621.984359] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] return func(*args, **kwargs) [ 621.984359] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.984359] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] raise e [ 621.984359] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.984359] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] nwinfo = self.network_api.allocate_for_instance( [ 621.984359] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 621.984359] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] created_port_ids = self._update_ports_for_instance( [ 621.984359] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 621.984359] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] with excutils.save_and_reraise_exception(): [ 621.984359] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.984723] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] self.force_reraise() [ 621.984723] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.984723] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] raise self.value [ 621.984723] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 621.984723] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] updated_port = self._update_port( [ 621.984723] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.984723] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] _ensure_no_port_binding_failure(port) [ 621.984723] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.984723] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] raise exception.PortBindingFailed(port_id=port['id']) [ 621.984723] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] nova.exception.PortBindingFailed: Binding failed for port fcb13275-a594-4d83-b081-729b1c23b331, please check neutron logs for more information. [ 621.984723] env[62692]: ERROR nova.compute.manager [instance: 549f672f-f21b-46b1-9002-42a634253624] [ 621.985014] env[62692]: DEBUG nova.compute.utils [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Binding failed for port fcb13275-a594-4d83-b081-729b1c23b331, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 621.988498] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.165s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.993077] env[62692]: INFO nova.compute.claims [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 622.003304] env[62692]: DEBUG nova.compute.manager [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Build of instance 549f672f-f21b-46b1-9002-42a634253624 was re-scheduled: Binding failed for port fcb13275-a594-4d83-b081-729b1c23b331, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 622.003304] env[62692]: DEBUG nova.compute.manager [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 622.003304] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "refresh_cache-549f672f-f21b-46b1-9002-42a634253624" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.003304] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquired lock "refresh_cache-549f672f-f21b-46b1-9002-42a634253624" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.003304] env[62692]: DEBUG nova.network.neutron [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 622.008018] env[62692]: DEBUG nova.compute.manager [req-aa4bb934-0611-45c1-8f92-3b26bfa9a05e req-5cb1f747-96d1-4ba1-84c6-e4946a058a8f service nova] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Received event network-changed-b344a4c1-5655-4faf-8f81-fa793eaf8eb6 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 622.008018] env[62692]: DEBUG nova.compute.manager [req-aa4bb934-0611-45c1-8f92-3b26bfa9a05e req-5cb1f747-96d1-4ba1-84c6-e4946a058a8f service nova] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Refreshing instance network info cache due to event network-changed-b344a4c1-5655-4faf-8f81-fa793eaf8eb6. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 622.008018] env[62692]: DEBUG oslo_concurrency.lockutils [req-aa4bb934-0611-45c1-8f92-3b26bfa9a05e req-5cb1f747-96d1-4ba1-84c6-e4946a058a8f service nova] Acquiring lock "refresh_cache-2d5a412a-66c0-46fc-84bc-2edb90f4593c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.149992] env[62692]: DEBUG nova.network.neutron [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.175522] env[62692]: DEBUG nova.compute.manager [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 622.189957] env[62692]: ERROR nova.compute.manager [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 466bf122-912c-490c-9871-bb2062bf6b3c, please check neutron logs for more information. [ 622.189957] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 622.189957] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 622.189957] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 622.189957] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 622.189957] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 622.189957] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 622.189957] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 622.189957] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.189957] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 622.189957] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.189957] env[62692]: ERROR nova.compute.manager raise self.value [ 622.189957] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 622.189957] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 622.189957] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.189957] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 622.190468] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.190468] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 622.190468] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 466bf122-912c-490c-9871-bb2062bf6b3c, please check neutron logs for more information. [ 622.190468] env[62692]: ERROR nova.compute.manager [ 622.190468] env[62692]: Traceback (most recent call last): [ 622.190468] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 622.190468] env[62692]: listener.cb(fileno) [ 622.190468] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 622.190468] env[62692]: result = function(*args, **kwargs) [ 622.190468] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 622.190468] env[62692]: return func(*args, **kwargs) [ 622.190468] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 622.190468] env[62692]: raise e [ 622.190468] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 622.190468] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 622.190468] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 622.190468] env[62692]: created_port_ids = self._update_ports_for_instance( [ 622.190468] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 622.190468] env[62692]: with excutils.save_and_reraise_exception(): [ 622.190468] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.190468] env[62692]: self.force_reraise() [ 622.190468] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.190468] env[62692]: raise self.value [ 622.190468] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 622.190468] env[62692]: updated_port = self._update_port( [ 622.190468] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.190468] env[62692]: _ensure_no_port_binding_failure(port) [ 622.190468] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.190468] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 622.191215] env[62692]: nova.exception.PortBindingFailed: Binding failed for port 466bf122-912c-490c-9871-bb2062bf6b3c, please check neutron logs for more information. [ 622.191215] env[62692]: Removing descriptor: 15 [ 622.191215] env[62692]: ERROR nova.compute.manager [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 466bf122-912c-490c-9871-bb2062bf6b3c, please check neutron logs for more information. [ 622.191215] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Traceback (most recent call last): [ 622.191215] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 622.191215] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] yield resources [ 622.191215] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 622.191215] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] self.driver.spawn(context, instance, image_meta, [ 622.191215] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 622.191215] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 622.191215] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 622.191215] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] vm_ref = self.build_virtual_machine(instance, [ 622.191585] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 622.191585] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] vif_infos = vmwarevif.get_vif_info(self._session, [ 622.191585] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 622.191585] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] for vif in network_info: [ 622.191585] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 622.191585] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] return self._sync_wrapper(fn, *args, **kwargs) [ 622.191585] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 622.191585] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] self.wait() [ 622.191585] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 622.191585] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] self[:] = self._gt.wait() [ 622.191585] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 622.191585] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] return self._exit_event.wait() [ 622.191585] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 622.191913] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] result = hub.switch() [ 622.191913] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 622.191913] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] return self.greenlet.switch() [ 622.191913] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 622.191913] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] result = function(*args, **kwargs) [ 622.191913] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 622.191913] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] return func(*args, **kwargs) [ 622.191913] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 622.191913] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] raise e [ 622.191913] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 622.191913] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] nwinfo = self.network_api.allocate_for_instance( [ 622.191913] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 622.191913] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] created_port_ids = self._update_ports_for_instance( [ 622.192265] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 622.192265] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] with excutils.save_and_reraise_exception(): [ 622.192265] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.192265] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] self.force_reraise() [ 622.192265] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.192265] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] raise self.value [ 622.192265] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 622.192265] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] updated_port = self._update_port( [ 622.192265] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.192265] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] _ensure_no_port_binding_failure(port) [ 622.192265] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.192265] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] raise exception.PortBindingFailed(port_id=port['id']) [ 622.192617] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] nova.exception.PortBindingFailed: Binding failed for port 466bf122-912c-490c-9871-bb2062bf6b3c, please check neutron logs for more information. [ 622.192617] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] [ 622.192617] env[62692]: INFO nova.compute.manager [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Terminating instance [ 622.195048] env[62692]: DEBUG oslo_concurrency.lockutils [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Acquiring lock "refresh_cache-331411c2-1758-46a4-ae08-8c575d0b8be2" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.201023] env[62692]: DEBUG oslo_concurrency.lockutils [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Acquired lock "refresh_cache-331411c2-1758-46a4-ae08-8c575d0b8be2" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.201023] env[62692]: DEBUG nova.network.neutron [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 622.203488] env[62692]: DEBUG nova.virt.hardware [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 622.203723] env[62692]: DEBUG nova.virt.hardware [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 622.203904] env[62692]: DEBUG nova.virt.hardware [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 622.204327] env[62692]: DEBUG nova.virt.hardware [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 622.204498] env[62692]: DEBUG nova.virt.hardware [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 622.204672] env[62692]: DEBUG nova.virt.hardware [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 622.204884] env[62692]: DEBUG nova.virt.hardware [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 622.205086] env[62692]: DEBUG nova.virt.hardware [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 622.205313] env[62692]: DEBUG nova.virt.hardware [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 622.205498] env[62692]: DEBUG nova.virt.hardware [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 622.205676] env[62692]: DEBUG nova.virt.hardware [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 622.207402] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03e2f95e-e577-4503-b583-09fcbd0618bc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.219900] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2689f8b-b7de-4f85-8582-a4809360e5c3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.304557] env[62692]: DEBUG nova.compute.manager [req-9d9a45da-6c25-46d7-9c9e-b070fc844269 req-04a23285-ecae-46d7-a907-f883d4c5e78b service nova] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Received event network-vif-deleted-b7e29e04-c6d2-4a03-bfdc-f406b818af64 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 622.342037] env[62692]: DEBUG oslo_concurrency.lockutils [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Acquiring lock "891170e7-35db-4faa-a1be-fcf376faee93" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.342665] env[62692]: DEBUG oslo_concurrency.lockutils [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Lock "891170e7-35db-4faa-a1be-fcf376faee93" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.344962] env[62692]: DEBUG nova.network.neutron [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Successfully created port: 1d2e6b57-3cfb-4d93-95e8-f855731915ed {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 622.541227] env[62692]: DEBUG nova.network.neutron [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 622.592918] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Acquiring lock "59882012-b01c-45ec-bd53-6e17e356ce5f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.593185] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Lock "59882012-b01c-45ec-bd53-6e17e356ce5f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.656859] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Releasing lock "refresh_cache-2d5a412a-66c0-46fc-84bc-2edb90f4593c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.657444] env[62692]: DEBUG nova.compute.manager [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 622.659594] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 622.659594] env[62692]: DEBUG oslo_concurrency.lockutils [req-aa4bb934-0611-45c1-8f92-3b26bfa9a05e req-5cb1f747-96d1-4ba1-84c6-e4946a058a8f service nova] Acquired lock "refresh_cache-2d5a412a-66c0-46fc-84bc-2edb90f4593c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.659594] env[62692]: DEBUG nova.network.neutron [req-aa4bb934-0611-45c1-8f92-3b26bfa9a05e req-5cb1f747-96d1-4ba1-84c6-e4946a058a8f service nova] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Refreshing network info cache for port b344a4c1-5655-4faf-8f81-fa793eaf8eb6 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 622.660862] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-069732f3-c29f-4150-ae2d-1d4ea9d7ddb4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.674330] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fa07075-7446-4318-bcb5-033e23480256 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.702670] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2d5a412a-66c0-46fc-84bc-2edb90f4593c could not be found. [ 622.702902] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 622.703095] env[62692]: INFO nova.compute.manager [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Took 0.05 seconds to destroy the instance on the hypervisor. [ 622.703809] env[62692]: DEBUG oslo.service.loopingcall [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 622.704289] env[62692]: DEBUG nova.network.neutron [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.705404] env[62692]: DEBUG nova.compute.manager [-] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 622.705506] env[62692]: DEBUG nova.network.neutron [-] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 622.750365] env[62692]: DEBUG nova.network.neutron [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 622.763062] env[62692]: DEBUG nova.network.neutron [-] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 622.947333] env[62692]: DEBUG nova.network.neutron [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.190536] env[62692]: DEBUG nova.network.neutron [req-aa4bb934-0611-45c1-8f92-3b26bfa9a05e req-5cb1f747-96d1-4ba1-84c6-e4946a058a8f service nova] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 623.210154] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Releasing lock "refresh_cache-549f672f-f21b-46b1-9002-42a634253624" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.212636] env[62692]: DEBUG nova.compute.manager [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 623.212636] env[62692]: DEBUG nova.compute.manager [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 623.212636] env[62692]: DEBUG nova.network.neutron [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 623.236646] env[62692]: DEBUG nova.network.neutron [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 623.241537] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f3a6260-355c-4ff6-975f-461cd01ecaa9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.250322] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b49d83e4-fc12-48ff-8eb7-192eb8befc42 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.287555] env[62692]: DEBUG nova.network.neutron [-] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.289822] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7349b0de-6eef-4394-b2ae-6b96f48b44c4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.300193] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6c9bd97-0b9d-46e6-ab3b-a15d41251a49 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.306228] env[62692]: DEBUG nova.network.neutron [req-aa4bb934-0611-45c1-8f92-3b26bfa9a05e req-5cb1f747-96d1-4ba1-84c6-e4946a058a8f service nova] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.320409] env[62692]: DEBUG nova.compute.provider_tree [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 623.449778] env[62692]: DEBUG oslo_concurrency.lockutils [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Releasing lock "refresh_cache-331411c2-1758-46a4-ae08-8c575d0b8be2" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.450343] env[62692]: DEBUG nova.compute.manager [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 623.450420] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 623.450689] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7bbb47c2-3a1c-40ff-9eee-79a682eec409 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.465825] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-240c23d8-2bb1-412e-8b21-d1a64b18ca4f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.491067] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 331411c2-1758-46a4-ae08-8c575d0b8be2 could not be found. [ 623.491401] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 623.491683] env[62692]: INFO nova.compute.manager [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 623.491960] env[62692]: DEBUG oslo.service.loopingcall [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 623.492322] env[62692]: DEBUG nova.compute.manager [-] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 623.492484] env[62692]: DEBUG nova.network.neutron [-] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 623.675754] env[62692]: DEBUG nova.network.neutron [-] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 623.738756] env[62692]: DEBUG nova.network.neutron [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.795942] env[62692]: INFO nova.compute.manager [-] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Took 1.09 seconds to deallocate network for instance. [ 623.798495] env[62692]: DEBUG nova.compute.claims [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 623.798671] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.809405] env[62692]: DEBUG oslo_concurrency.lockutils [req-aa4bb934-0611-45c1-8f92-3b26bfa9a05e req-5cb1f747-96d1-4ba1-84c6-e4946a058a8f service nova] Releasing lock "refresh_cache-2d5a412a-66c0-46fc-84bc-2edb90f4593c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.823050] env[62692]: DEBUG nova.scheduler.client.report [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 624.181020] env[62692]: DEBUG nova.network.neutron [-] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.242146] env[62692]: INFO nova.compute.manager [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 549f672f-f21b-46b1-9002-42a634253624] Took 1.03 seconds to deallocate network for instance. [ 624.329585] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.340s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.329585] env[62692]: DEBUG nova.compute.manager [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 624.334184] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.254s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.334384] env[62692]: INFO nova.compute.claims [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 624.681625] env[62692]: INFO nova.compute.manager [-] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Took 1.19 seconds to deallocate network for instance. [ 624.684515] env[62692]: DEBUG nova.compute.claims [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 624.684798] env[62692]: DEBUG oslo_concurrency.lockutils [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.839945] env[62692]: DEBUG nova.compute.utils [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 624.843793] env[62692]: DEBUG nova.compute.manager [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 624.846232] env[62692]: DEBUG nova.network.neutron [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 625.015512] env[62692]: DEBUG nova.policy [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fd47001f41914566b331ed8fff927f8b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dc5079f79f3048c39d3d88b7038ee206', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 625.264182] env[62692]: DEBUG nova.compute.manager [req-d07665fa-39e5-4675-b283-59669f3251f9 req-9dbde2fe-3f5a-47ed-b233-fbecae08a077 service nova] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Received event network-vif-deleted-b344a4c1-5655-4faf-8f81-fa793eaf8eb6 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 625.290730] env[62692]: INFO nova.scheduler.client.report [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Deleted allocations for instance 549f672f-f21b-46b1-9002-42a634253624 [ 625.309325] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Acquiring lock "b014c45b-4d27-4ade-b823-0dd64dfea2e2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.311839] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Lock "b014c45b-4d27-4ade-b823-0dd64dfea2e2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.345228] env[62692]: DEBUG nova.compute.manager [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 625.593455] env[62692]: DEBUG nova.compute.manager [req-0b264418-e300-4ba5-912b-a0bc3bf1c22e req-134eeff2-6ca9-4a09-a49c-1229a049a856 service nova] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Received event network-changed-466bf122-912c-490c-9871-bb2062bf6b3c {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 625.593455] env[62692]: DEBUG nova.compute.manager [req-0b264418-e300-4ba5-912b-a0bc3bf1c22e req-134eeff2-6ca9-4a09-a49c-1229a049a856 service nova] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Refreshing instance network info cache due to event network-changed-466bf122-912c-490c-9871-bb2062bf6b3c. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 625.593455] env[62692]: DEBUG oslo_concurrency.lockutils [req-0b264418-e300-4ba5-912b-a0bc3bf1c22e req-134eeff2-6ca9-4a09-a49c-1229a049a856 service nova] Acquiring lock "refresh_cache-331411c2-1758-46a4-ae08-8c575d0b8be2" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.593455] env[62692]: DEBUG oslo_concurrency.lockutils [req-0b264418-e300-4ba5-912b-a0bc3bf1c22e req-134eeff2-6ca9-4a09-a49c-1229a049a856 service nova] Acquired lock "refresh_cache-331411c2-1758-46a4-ae08-8c575d0b8be2" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.593455] env[62692]: DEBUG nova.network.neutron [req-0b264418-e300-4ba5-912b-a0bc3bf1c22e req-134eeff2-6ca9-4a09-a49c-1229a049a856 service nova] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Refreshing network info cache for port 466bf122-912c-490c-9871-bb2062bf6b3c {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 625.642103] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdfd248d-0148-4d58-85c2-dad08ec30fc4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.650956] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cce40c60-4c70-4a39-b736-9841cd995d96 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.656438] env[62692]: ERROR nova.compute.manager [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1d2e6b57-3cfb-4d93-95e8-f855731915ed, please check neutron logs for more information. [ 625.656438] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 625.656438] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.656438] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 625.656438] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 625.656438] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 625.656438] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 625.656438] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 625.656438] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.656438] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 625.656438] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.656438] env[62692]: ERROR nova.compute.manager raise self.value [ 625.656438] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 625.656438] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 625.656438] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.656438] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 625.656936] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.656936] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 625.656936] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1d2e6b57-3cfb-4d93-95e8-f855731915ed, please check neutron logs for more information. [ 625.656936] env[62692]: ERROR nova.compute.manager [ 625.656936] env[62692]: Traceback (most recent call last): [ 625.656936] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 625.656936] env[62692]: listener.cb(fileno) [ 625.656936] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.656936] env[62692]: result = function(*args, **kwargs) [ 625.656936] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 625.656936] env[62692]: return func(*args, **kwargs) [ 625.656936] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 625.656936] env[62692]: raise e [ 625.656936] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.656936] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 625.656936] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 625.656936] env[62692]: created_port_ids = self._update_ports_for_instance( [ 625.656936] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 625.656936] env[62692]: with excutils.save_and_reraise_exception(): [ 625.656936] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.656936] env[62692]: self.force_reraise() [ 625.656936] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.656936] env[62692]: raise self.value [ 625.656936] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 625.656936] env[62692]: updated_port = self._update_port( [ 625.656936] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.656936] env[62692]: _ensure_no_port_binding_failure(port) [ 625.656936] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.656936] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 625.657791] env[62692]: nova.exception.PortBindingFailed: Binding failed for port 1d2e6b57-3cfb-4d93-95e8-f855731915ed, please check neutron logs for more information. [ 625.657791] env[62692]: Removing descriptor: 19 [ 625.657791] env[62692]: ERROR nova.compute.manager [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1d2e6b57-3cfb-4d93-95e8-f855731915ed, please check neutron logs for more information. [ 625.657791] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Traceback (most recent call last): [ 625.657791] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 625.657791] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] yield resources [ 625.657791] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 625.657791] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] self.driver.spawn(context, instance, image_meta, [ 625.657791] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 625.657791] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 625.657791] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 625.657791] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] vm_ref = self.build_virtual_machine(instance, [ 625.658205] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 625.658205] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] vif_infos = vmwarevif.get_vif_info(self._session, [ 625.658205] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 625.658205] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] for vif in network_info: [ 625.658205] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 625.658205] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] return self._sync_wrapper(fn, *args, **kwargs) [ 625.658205] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 625.658205] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] self.wait() [ 625.658205] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 625.658205] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] self[:] = self._gt.wait() [ 625.658205] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 625.658205] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] return self._exit_event.wait() [ 625.658205] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 625.658588] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] result = hub.switch() [ 625.658588] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 625.658588] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] return self.greenlet.switch() [ 625.658588] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.658588] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] result = function(*args, **kwargs) [ 625.658588] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 625.658588] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] return func(*args, **kwargs) [ 625.658588] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 625.658588] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] raise e [ 625.658588] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.658588] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] nwinfo = self.network_api.allocate_for_instance( [ 625.658588] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 625.658588] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] created_port_ids = self._update_ports_for_instance( [ 625.658936] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 625.658936] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] with excutils.save_and_reraise_exception(): [ 625.658936] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.658936] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] self.force_reraise() [ 625.658936] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.658936] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] raise self.value [ 625.658936] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 625.658936] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] updated_port = self._update_port( [ 625.658936] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.658936] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] _ensure_no_port_binding_failure(port) [ 625.658936] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.658936] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] raise exception.PortBindingFailed(port_id=port['id']) [ 625.659298] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] nova.exception.PortBindingFailed: Binding failed for port 1d2e6b57-3cfb-4d93-95e8-f855731915ed, please check neutron logs for more information. [ 625.659298] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] [ 625.659298] env[62692]: INFO nova.compute.manager [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Terminating instance [ 625.660053] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Acquiring lock "refresh_cache-5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.660214] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Acquired lock "refresh_cache-5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.660376] env[62692]: DEBUG nova.network.neutron [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 625.691507] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1275df92-1f7b-49ce-8833-04c0b934481a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.702442] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b581993-dd62-4a7f-a295-15318d0816b0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.719063] env[62692]: DEBUG nova.compute.provider_tree [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 625.743620] env[62692]: DEBUG nova.network.neutron [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 625.798916] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a7924186-b379-4e29-a3a6-8e05a5d40fba tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "549f672f-f21b-46b1-9002-42a634253624" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.763s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.801039] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "549f672f-f21b-46b1-9002-42a634253624" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 16.569s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.801039] env[62692]: INFO nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 549f672f-f21b-46b1-9002-42a634253624] During sync_power_state the instance has a pending task (spawning). Skip. [ 625.801039] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "549f672f-f21b-46b1-9002-42a634253624" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.891125] env[62692]: DEBUG nova.network.neutron [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Successfully created port: d33d9daf-8548-4f1d-a09e-ed9348cf5d0c {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 626.085482] env[62692]: DEBUG nova.network.neutron [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.126822] env[62692]: DEBUG nova.network.neutron [req-0b264418-e300-4ba5-912b-a0bc3bf1c22e req-134eeff2-6ca9-4a09-a49c-1229a049a856 service nova] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 626.203834] env[62692]: DEBUG nova.network.neutron [req-0b264418-e300-4ba5-912b-a0bc3bf1c22e req-134eeff2-6ca9-4a09-a49c-1229a049a856 service nova] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.229462] env[62692]: DEBUG nova.scheduler.client.report [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 626.305374] env[62692]: DEBUG nova.compute.manager [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 626.364160] env[62692]: DEBUG nova.compute.manager [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 626.402126] env[62692]: DEBUG nova.virt.hardware [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 626.402284] env[62692]: DEBUG nova.virt.hardware [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 626.402445] env[62692]: DEBUG nova.virt.hardware [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 626.402623] env[62692]: DEBUG nova.virt.hardware [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 626.402786] env[62692]: DEBUG nova.virt.hardware [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 626.402909] env[62692]: DEBUG nova.virt.hardware [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 626.403159] env[62692]: DEBUG nova.virt.hardware [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 626.403354] env[62692]: DEBUG nova.virt.hardware [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 626.403592] env[62692]: DEBUG nova.virt.hardware [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 626.403690] env[62692]: DEBUG nova.virt.hardware [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 626.403846] env[62692]: DEBUG nova.virt.hardware [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 626.405040] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dcae4f5-ddd3-4984-8fa8-19976c1cef26 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.414660] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a986c96b-d79a-4020-9105-e54110497fa9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.587845] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Releasing lock "refresh_cache-5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.588315] env[62692]: DEBUG nova.compute.manager [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 626.591237] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 626.591573] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e169f7ea-0192-48c3-ab0a-9f7a0e6af6d5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.600978] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-419ff284-989b-468d-b3a7-034ddddf2f10 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.633466] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e could not be found. [ 626.633834] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 626.634108] env[62692]: INFO nova.compute.manager [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Took 0.05 seconds to destroy the instance on the hypervisor. [ 626.634394] env[62692]: DEBUG oslo.service.loopingcall [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 626.634633] env[62692]: DEBUG nova.compute.manager [-] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 626.634765] env[62692]: DEBUG nova.network.neutron [-] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 626.655437] env[62692]: DEBUG nova.network.neutron [-] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 626.711056] env[62692]: DEBUG oslo_concurrency.lockutils [req-0b264418-e300-4ba5-912b-a0bc3bf1c22e req-134eeff2-6ca9-4a09-a49c-1229a049a856 service nova] Releasing lock "refresh_cache-331411c2-1758-46a4-ae08-8c575d0b8be2" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.711959] env[62692]: DEBUG nova.compute.manager [req-0b264418-e300-4ba5-912b-a0bc3bf1c22e req-134eeff2-6ca9-4a09-a49c-1229a049a856 service nova] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Received event network-vif-deleted-466bf122-912c-490c-9871-bb2062bf6b3c {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 626.739251] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.403s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.739251] env[62692]: DEBUG nova.compute.manager [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 626.740574] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.648s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.839029] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.949223] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Acquiring lock "6883c19f-1e3a-4eac-be74-8e725faee77a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.951712] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Lock "6883c19f-1e3a-4eac-be74-8e725faee77a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.003s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.157591] env[62692]: DEBUG nova.network.neutron [-] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.249288] env[62692]: DEBUG nova.compute.utils [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 627.265786] env[62692]: DEBUG nova.compute.manager [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 627.265786] env[62692]: DEBUG nova.network.neutron [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 627.384703] env[62692]: DEBUG nova.policy [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8adae7ddecbc4aa28694a228f52b5bd6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b00f181776484ecebc81d7469d81153f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 627.535751] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb8c6ee0-a069-4831-a51f-8a157cef82c8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.545448] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b934e8-d5a1-467c-8fe6-277e95065f63 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.586657] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6be623d3-82ed-4824-b19f-39991946a6d5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.597302] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-799a16aa-fa7f-4c46-ac47-79e17c306b8c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.610557] env[62692]: DEBUG nova.compute.provider_tree [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 627.662609] env[62692]: INFO nova.compute.manager [-] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Took 1.03 seconds to deallocate network for instance. [ 627.666110] env[62692]: DEBUG nova.compute.claims [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 627.666282] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.734012] env[62692]: ERROR nova.compute.manager [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d33d9daf-8548-4f1d-a09e-ed9348cf5d0c, please check neutron logs for more information. [ 627.734012] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 627.734012] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.734012] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 627.734012] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 627.734012] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 627.734012] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 627.734012] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 627.734012] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.734012] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 627.734012] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.734012] env[62692]: ERROR nova.compute.manager raise self.value [ 627.734012] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 627.734012] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 627.734012] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.734012] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 627.734529] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.734529] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 627.734529] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d33d9daf-8548-4f1d-a09e-ed9348cf5d0c, please check neutron logs for more information. [ 627.734529] env[62692]: ERROR nova.compute.manager [ 627.734529] env[62692]: Traceback (most recent call last): [ 627.734529] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 627.734529] env[62692]: listener.cb(fileno) [ 627.734529] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.734529] env[62692]: result = function(*args, **kwargs) [ 627.734529] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 627.734529] env[62692]: return func(*args, **kwargs) [ 627.734529] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 627.734529] env[62692]: raise e [ 627.734529] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.734529] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 627.734529] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 627.734529] env[62692]: created_port_ids = self._update_ports_for_instance( [ 627.734529] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 627.734529] env[62692]: with excutils.save_and_reraise_exception(): [ 627.734529] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.734529] env[62692]: self.force_reraise() [ 627.734529] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.734529] env[62692]: raise self.value [ 627.734529] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 627.734529] env[62692]: updated_port = self._update_port( [ 627.734529] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.734529] env[62692]: _ensure_no_port_binding_failure(port) [ 627.734529] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.734529] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 627.735436] env[62692]: nova.exception.PortBindingFailed: Binding failed for port d33d9daf-8548-4f1d-a09e-ed9348cf5d0c, please check neutron logs for more information. [ 627.735436] env[62692]: Removing descriptor: 17 [ 627.735436] env[62692]: ERROR nova.compute.manager [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d33d9daf-8548-4f1d-a09e-ed9348cf5d0c, please check neutron logs for more information. [ 627.735436] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] Traceback (most recent call last): [ 627.735436] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 627.735436] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] yield resources [ 627.735436] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 627.735436] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] self.driver.spawn(context, instance, image_meta, [ 627.735436] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 627.735436] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] self._vmops.spawn(context, instance, image_meta, injected_files, [ 627.735436] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 627.735436] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] vm_ref = self.build_virtual_machine(instance, [ 627.735834] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 627.735834] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] vif_infos = vmwarevif.get_vif_info(self._session, [ 627.735834] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 627.735834] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] for vif in network_info: [ 627.735834] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 627.735834] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] return self._sync_wrapper(fn, *args, **kwargs) [ 627.735834] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 627.735834] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] self.wait() [ 627.735834] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 627.735834] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] self[:] = self._gt.wait() [ 627.735834] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 627.735834] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] return self._exit_event.wait() [ 627.735834] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 627.736535] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] result = hub.switch() [ 627.736535] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 627.736535] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] return self.greenlet.switch() [ 627.736535] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.736535] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] result = function(*args, **kwargs) [ 627.736535] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 627.736535] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] return func(*args, **kwargs) [ 627.736535] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 627.736535] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] raise e [ 627.736535] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.736535] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] nwinfo = self.network_api.allocate_for_instance( [ 627.736535] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 627.736535] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] created_port_ids = self._update_ports_for_instance( [ 627.737054] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 627.737054] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] with excutils.save_and_reraise_exception(): [ 627.737054] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.737054] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] self.force_reraise() [ 627.737054] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.737054] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] raise self.value [ 627.737054] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 627.737054] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] updated_port = self._update_port( [ 627.737054] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.737054] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] _ensure_no_port_binding_failure(port) [ 627.737054] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.737054] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] raise exception.PortBindingFailed(port_id=port['id']) [ 627.737470] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] nova.exception.PortBindingFailed: Binding failed for port d33d9daf-8548-4f1d-a09e-ed9348cf5d0c, please check neutron logs for more information. [ 627.737470] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] [ 627.737470] env[62692]: INFO nova.compute.manager [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Terminating instance [ 627.740551] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Acquiring lock "refresh_cache-647d9514-5c31-4970-82e7-626e08f00237" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.740551] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Acquired lock "refresh_cache-647d9514-5c31-4970-82e7-626e08f00237" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.740551] env[62692]: DEBUG nova.network.neutron [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 627.762134] env[62692]: DEBUG nova.compute.manager [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 628.113947] env[62692]: DEBUG nova.scheduler.client.report [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 628.237766] env[62692]: DEBUG nova.compute.manager [req-026bc9dc-0247-42c8-9a74-59deddfc5405 req-7c10edae-383b-4c9e-ab6e-75b047512610 service nova] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Received event network-changed-1d2e6b57-3cfb-4d93-95e8-f855731915ed {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 628.238290] env[62692]: DEBUG nova.compute.manager [req-026bc9dc-0247-42c8-9a74-59deddfc5405 req-7c10edae-383b-4c9e-ab6e-75b047512610 service nova] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Refreshing instance network info cache due to event network-changed-1d2e6b57-3cfb-4d93-95e8-f855731915ed. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 628.240626] env[62692]: DEBUG oslo_concurrency.lockutils [req-026bc9dc-0247-42c8-9a74-59deddfc5405 req-7c10edae-383b-4c9e-ab6e-75b047512610 service nova] Acquiring lock "refresh_cache-5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.240626] env[62692]: DEBUG oslo_concurrency.lockutils [req-026bc9dc-0247-42c8-9a74-59deddfc5405 req-7c10edae-383b-4c9e-ab6e-75b047512610 service nova] Acquired lock "refresh_cache-5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.240626] env[62692]: DEBUG nova.network.neutron [req-026bc9dc-0247-42c8-9a74-59deddfc5405 req-7c10edae-383b-4c9e-ab6e-75b047512610 service nova] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Refreshing network info cache for port 1d2e6b57-3cfb-4d93-95e8-f855731915ed {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 628.327150] env[62692]: DEBUG nova.network.neutron [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 628.360925] env[62692]: DEBUG nova.network.neutron [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Successfully created port: dc5d38ec-06fc-440f-bd69-0a197ea6da73 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 628.618831] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.878s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.619166] env[62692]: ERROR nova.compute.manager [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4312822f-3bf3-49ed-864c-fd84521b74ee, please check neutron logs for more information. [ 628.619166] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Traceback (most recent call last): [ 628.619166] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 628.619166] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] self.driver.spawn(context, instance, image_meta, [ 628.619166] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 628.619166] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 628.619166] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 628.619166] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] vm_ref = self.build_virtual_machine(instance, [ 628.619166] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 628.619166] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] vif_infos = vmwarevif.get_vif_info(self._session, [ 628.619166] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 628.619478] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] for vif in network_info: [ 628.619478] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 628.619478] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] return self._sync_wrapper(fn, *args, **kwargs) [ 628.619478] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 628.619478] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] self.wait() [ 628.619478] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 628.619478] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] self[:] = self._gt.wait() [ 628.619478] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 628.619478] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] return self._exit_event.wait() [ 628.619478] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 628.619478] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] result = hub.switch() [ 628.619478] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 628.619478] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] return self.greenlet.switch() [ 628.619828] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.619828] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] result = function(*args, **kwargs) [ 628.619828] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 628.619828] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] return func(*args, **kwargs) [ 628.619828] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 628.619828] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] raise e [ 628.619828] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.619828] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] nwinfo = self.network_api.allocate_for_instance( [ 628.619828] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 628.619828] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] created_port_ids = self._update_ports_for_instance( [ 628.619828] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 628.619828] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] with excutils.save_and_reraise_exception(): [ 628.619828] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.620222] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] self.force_reraise() [ 628.620222] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.620222] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] raise self.value [ 628.620222] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 628.620222] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] updated_port = self._update_port( [ 628.620222] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.620222] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] _ensure_no_port_binding_failure(port) [ 628.620222] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.620222] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] raise exception.PortBindingFailed(port_id=port['id']) [ 628.620222] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] nova.exception.PortBindingFailed: Binding failed for port 4312822f-3bf3-49ed-864c-fd84521b74ee, please check neutron logs for more information. [ 628.620222] env[62692]: ERROR nova.compute.manager [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] [ 628.620523] env[62692]: DEBUG nova.compute.utils [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Binding failed for port 4312822f-3bf3-49ed-864c-fd84521b74ee, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 628.621952] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.221s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.624658] env[62692]: DEBUG nova.compute.manager [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Build of instance 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6 was re-scheduled: Binding failed for port 4312822f-3bf3-49ed-864c-fd84521b74ee, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 628.624914] env[62692]: DEBUG nova.compute.manager [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 628.625537] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Acquiring lock "refresh_cache-7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.625537] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Acquired lock "refresh_cache-7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.625537] env[62692]: DEBUG nova.network.neutron [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 628.668972] env[62692]: DEBUG nova.network.neutron [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.778420] env[62692]: DEBUG nova.compute.manager [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 628.783621] env[62692]: DEBUG nova.network.neutron [req-026bc9dc-0247-42c8-9a74-59deddfc5405 req-7c10edae-383b-4c9e-ab6e-75b047512610 service nova] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 628.811404] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Acquiring lock "844b3803-8a01-484e-8908-dbae27188acb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.811633] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Lock "844b3803-8a01-484e-8908-dbae27188acb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.815694] env[62692]: DEBUG nova.virt.hardware [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 628.815812] env[62692]: DEBUG nova.virt.hardware [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 628.815964] env[62692]: DEBUG nova.virt.hardware [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 628.816168] env[62692]: DEBUG nova.virt.hardware [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 628.816318] env[62692]: DEBUG nova.virt.hardware [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 628.816464] env[62692]: DEBUG nova.virt.hardware [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 628.816678] env[62692]: DEBUG nova.virt.hardware [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 628.816910] env[62692]: DEBUG nova.virt.hardware [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 628.817110] env[62692]: DEBUG nova.virt.hardware [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 628.817277] env[62692]: DEBUG nova.virt.hardware [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 628.817449] env[62692]: DEBUG nova.virt.hardware [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 628.818750] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8df39e36-dcf7-461e-a5e3-110d6042c6ef {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.827389] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b34baddb-0e1c-4303-9d09-c53a932f251e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.171675] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Releasing lock "refresh_cache-647d9514-5c31-4970-82e7-626e08f00237" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.172165] env[62692]: DEBUG nova.compute.manager [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 629.173145] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 629.177302] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-742258d2-31c1-4d4b-83ce-81b3b2fdca60 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.183650] env[62692]: DEBUG nova.network.neutron [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 629.192674] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2bdf33-1e5a-4b0f-978d-efa0cb8d421a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.224051] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 647d9514-5c31-4970-82e7-626e08f00237 could not be found. [ 629.224366] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 629.224564] env[62692]: INFO nova.compute.manager [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Took 0.05 seconds to destroy the instance on the hypervisor. [ 629.224886] env[62692]: DEBUG oslo.service.loopingcall [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 629.227256] env[62692]: DEBUG nova.compute.manager [-] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 629.227474] env[62692]: DEBUG nova.network.neutron [-] [instance: 647d9514-5c31-4970-82e7-626e08f00237] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 629.289020] env[62692]: DEBUG nova.network.neutron [-] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 629.334234] env[62692]: DEBUG nova.network.neutron [req-026bc9dc-0247-42c8-9a74-59deddfc5405 req-7c10edae-383b-4c9e-ab6e-75b047512610 service nova] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.449208] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05b81a24-8766-47c0-9d65-28650d4b4213 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.456878] env[62692]: DEBUG nova.network.neutron [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.462501] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9340a1bf-eb36-4664-9895-c81eba601b70 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.501328] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40d99b5c-2074-4b43-afe0-89ad09b245c9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.510700] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d40e748c-bc6f-4561-b16e-93c7102d56a8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.533488] env[62692]: DEBUG nova.compute.provider_tree [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 629.792739] env[62692]: DEBUG nova.network.neutron [-] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.838373] env[62692]: DEBUG oslo_concurrency.lockutils [req-026bc9dc-0247-42c8-9a74-59deddfc5405 req-7c10edae-383b-4c9e-ab6e-75b047512610 service nova] Releasing lock "refresh_cache-5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.838637] env[62692]: DEBUG nova.compute.manager [req-026bc9dc-0247-42c8-9a74-59deddfc5405 req-7c10edae-383b-4c9e-ab6e-75b047512610 service nova] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Received event network-vif-deleted-1d2e6b57-3cfb-4d93-95e8-f855731915ed {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 629.838830] env[62692]: DEBUG nova.compute.manager [req-026bc9dc-0247-42c8-9a74-59deddfc5405 req-7c10edae-383b-4c9e-ab6e-75b047512610 service nova] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Received event network-changed-d33d9daf-8548-4f1d-a09e-ed9348cf5d0c {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 629.838995] env[62692]: DEBUG nova.compute.manager [req-026bc9dc-0247-42c8-9a74-59deddfc5405 req-7c10edae-383b-4c9e-ab6e-75b047512610 service nova] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Refreshing instance network info cache due to event network-changed-d33d9daf-8548-4f1d-a09e-ed9348cf5d0c. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 629.839214] env[62692]: DEBUG oslo_concurrency.lockutils [req-026bc9dc-0247-42c8-9a74-59deddfc5405 req-7c10edae-383b-4c9e-ab6e-75b047512610 service nova] Acquiring lock "refresh_cache-647d9514-5c31-4970-82e7-626e08f00237" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.839349] env[62692]: DEBUG oslo_concurrency.lockutils [req-026bc9dc-0247-42c8-9a74-59deddfc5405 req-7c10edae-383b-4c9e-ab6e-75b047512610 service nova] Acquired lock "refresh_cache-647d9514-5c31-4970-82e7-626e08f00237" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.839609] env[62692]: DEBUG nova.network.neutron [req-026bc9dc-0247-42c8-9a74-59deddfc5405 req-7c10edae-383b-4c9e-ab6e-75b047512610 service nova] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Refreshing network info cache for port d33d9daf-8548-4f1d-a09e-ed9348cf5d0c {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 629.936350] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Acquiring lock "0aac5bf9-2af4-4745-a961-46986fb57779" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.936350] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Lock "0aac5bf9-2af4-4745-a961-46986fb57779" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.962704] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Releasing lock "refresh_cache-7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.962704] env[62692]: DEBUG nova.compute.manager [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 629.962704] env[62692]: DEBUG nova.compute.manager [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 629.962704] env[62692]: DEBUG nova.network.neutron [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 629.995418] env[62692]: DEBUG nova.network.neutron [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 630.038597] env[62692]: DEBUG nova.scheduler.client.report [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 630.296198] env[62692]: INFO nova.compute.manager [-] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Took 1.07 seconds to deallocate network for instance. [ 630.298996] env[62692]: DEBUG nova.compute.claims [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 630.298996] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.381368] env[62692]: DEBUG nova.network.neutron [req-026bc9dc-0247-42c8-9a74-59deddfc5405 req-7c10edae-383b-4c9e-ab6e-75b047512610 service nova] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 630.499367] env[62692]: DEBUG nova.network.neutron [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.545893] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.922s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.545893] env[62692]: ERROR nova.compute.manager [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8a34d418-e31d-40da-b68f-889fef817fb8, please check neutron logs for more information. [ 630.545893] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Traceback (most recent call last): [ 630.545893] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 630.545893] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] self.driver.spawn(context, instance, image_meta, [ 630.545893] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 630.545893] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 630.545893] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 630.545893] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] vm_ref = self.build_virtual_machine(instance, [ 630.546254] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 630.546254] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] vif_infos = vmwarevif.get_vif_info(self._session, [ 630.546254] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 630.546254] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] for vif in network_info: [ 630.546254] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 630.546254] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] return self._sync_wrapper(fn, *args, **kwargs) [ 630.546254] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 630.546254] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] self.wait() [ 630.546254] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 630.546254] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] self[:] = self._gt.wait() [ 630.546254] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 630.546254] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] return self._exit_event.wait() [ 630.546254] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 630.546603] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] result = hub.switch() [ 630.546603] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 630.546603] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] return self.greenlet.switch() [ 630.546603] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 630.546603] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] result = function(*args, **kwargs) [ 630.546603] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 630.546603] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] return func(*args, **kwargs) [ 630.546603] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 630.546603] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] raise e [ 630.546603] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.546603] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] nwinfo = self.network_api.allocate_for_instance( [ 630.546603] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 630.546603] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] created_port_ids = self._update_ports_for_instance( [ 630.547036] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 630.547036] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] with excutils.save_and_reraise_exception(): [ 630.547036] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.547036] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] self.force_reraise() [ 630.547036] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.547036] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] raise self.value [ 630.547036] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 630.547036] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] updated_port = self._update_port( [ 630.547036] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.547036] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] _ensure_no_port_binding_failure(port) [ 630.547036] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.547036] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] raise exception.PortBindingFailed(port_id=port['id']) [ 630.547344] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] nova.exception.PortBindingFailed: Binding failed for port 8a34d418-e31d-40da-b68f-889fef817fb8, please check neutron logs for more information. [ 630.547344] env[62692]: ERROR nova.compute.manager [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] [ 630.547344] env[62692]: DEBUG nova.compute.utils [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Binding failed for port 8a34d418-e31d-40da-b68f-889fef817fb8, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 630.548315] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.088s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.549877] env[62692]: INFO nova.compute.claims [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 630.553087] env[62692]: DEBUG nova.compute.manager [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Build of instance a16b5c50-a48a-4dfb-b2cb-bfe3723d108e was re-scheduled: Binding failed for port 8a34d418-e31d-40da-b68f-889fef817fb8, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 630.553568] env[62692]: DEBUG nova.compute.manager [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 630.553797] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "refresh_cache-a16b5c50-a48a-4dfb-b2cb-bfe3723d108e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.553941] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquired lock "refresh_cache-a16b5c50-a48a-4dfb-b2cb-bfe3723d108e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.554115] env[62692]: DEBUG nova.network.neutron [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 630.675516] env[62692]: DEBUG nova.network.neutron [req-026bc9dc-0247-42c8-9a74-59deddfc5405 req-7c10edae-383b-4c9e-ab6e-75b047512610 service nova] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.892164] env[62692]: DEBUG nova.compute.manager [req-579929f3-7154-4e67-a508-e6eb509184c7 req-307ce7b4-a3ca-426f-b868-cc180db774eb service nova] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Received event network-vif-deleted-d33d9daf-8548-4f1d-a09e-ed9348cf5d0c {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 631.004544] env[62692]: INFO nova.compute.manager [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] Took 1.04 seconds to deallocate network for instance. [ 631.114016] env[62692]: DEBUG nova.network.neutron [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 631.178882] env[62692]: DEBUG oslo_concurrency.lockutils [req-026bc9dc-0247-42c8-9a74-59deddfc5405 req-7c10edae-383b-4c9e-ab6e-75b047512610 service nova] Releasing lock "refresh_cache-647d9514-5c31-4970-82e7-626e08f00237" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.435332] env[62692]: DEBUG nova.network.neutron [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.903482] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33be164f-aa52-4a67-bc05-9925f4a479c9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.914141] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7d6baa7-bd6d-4082-bae5-e9e4f7adde6a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.948657] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Releasing lock "refresh_cache-a16b5c50-a48a-4dfb-b2cb-bfe3723d108e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.948967] env[62692]: DEBUG nova.compute.manager [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 631.949293] env[62692]: DEBUG nova.compute.manager [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 631.949505] env[62692]: DEBUG nova.network.neutron [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 631.953830] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aabab74a-49c0-4029-b84c-62545f7d88af {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.959868] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba94c400-af63-4e8f-8f7c-89c74bd6a392 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.975016] env[62692]: DEBUG nova.compute.provider_tree [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.011973] env[62692]: DEBUG nova.network.neutron [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 632.044725] env[62692]: INFO nova.scheduler.client.report [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Deleted allocations for instance 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6 [ 632.477781] env[62692]: DEBUG nova.scheduler.client.report [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 632.517326] env[62692]: DEBUG nova.network.neutron [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.559405] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a3d495e8-7827-4068-b1b2-8e392113ccff tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Lock "7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.705s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.562934] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 23.331s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.562934] env[62692]: INFO nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6] During sync_power_state the instance has a pending task (networking). Skip. [ 632.562934] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "7a7f088b-00d8-41ab-8ffe-f5a12e3bb5d6" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.988502] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.440s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.989156] env[62692]: DEBUG nova.compute.manager [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 632.996022] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.775s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.996022] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.996022] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62692) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 632.996022] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.833s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.997018] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7816168-2aa8-4ed4-99e2-92831882e050 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.006583] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82d32555-02a2-439f-a787-0607212c7c39 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.025785] env[62692]: INFO nova.compute.manager [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] Took 1.08 seconds to deallocate network for instance. [ 633.030290] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47126ddc-8ca5-4b09-8243-a05cc84282d6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.037797] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06ba478b-abf9-4c6a-8d75-07bdaca3148c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.076143] env[62692]: DEBUG nova.compute.manager [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 633.079331] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181526MB free_disk=167GB free_vcpus=48 pci_devices=None {{(pid=62692) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 633.079331] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.498415] env[62692]: DEBUG nova.compute.utils [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 633.506053] env[62692]: DEBUG nova.compute.manager [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Not allocating networking since 'none' was specified. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 633.562854] env[62692]: ERROR nova.compute.manager [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port dc5d38ec-06fc-440f-bd69-0a197ea6da73, please check neutron logs for more information. [ 633.562854] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 633.562854] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.562854] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 633.562854] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 633.562854] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 633.562854] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 633.562854] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 633.562854] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.562854] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 633.562854] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.562854] env[62692]: ERROR nova.compute.manager raise self.value [ 633.562854] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 633.562854] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 633.562854] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.562854] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 633.563714] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.563714] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 633.563714] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port dc5d38ec-06fc-440f-bd69-0a197ea6da73, please check neutron logs for more information. [ 633.563714] env[62692]: ERROR nova.compute.manager [ 633.563714] env[62692]: Traceback (most recent call last): [ 633.563714] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 633.563714] env[62692]: listener.cb(fileno) [ 633.563714] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.563714] env[62692]: result = function(*args, **kwargs) [ 633.563714] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 633.563714] env[62692]: return func(*args, **kwargs) [ 633.563714] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 633.563714] env[62692]: raise e [ 633.563714] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.563714] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 633.563714] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 633.563714] env[62692]: created_port_ids = self._update_ports_for_instance( [ 633.563714] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 633.563714] env[62692]: with excutils.save_and_reraise_exception(): [ 633.563714] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.563714] env[62692]: self.force_reraise() [ 633.563714] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.563714] env[62692]: raise self.value [ 633.563714] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 633.563714] env[62692]: updated_port = self._update_port( [ 633.563714] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.563714] env[62692]: _ensure_no_port_binding_failure(port) [ 633.563714] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.563714] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 633.564789] env[62692]: nova.exception.PortBindingFailed: Binding failed for port dc5d38ec-06fc-440f-bd69-0a197ea6da73, please check neutron logs for more information. [ 633.564789] env[62692]: Removing descriptor: 19 [ 633.564789] env[62692]: ERROR nova.compute.manager [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port dc5d38ec-06fc-440f-bd69-0a197ea6da73, please check neutron logs for more information. [ 633.564789] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Traceback (most recent call last): [ 633.564789] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 633.564789] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] yield resources [ 633.564789] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 633.564789] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] self.driver.spawn(context, instance, image_meta, [ 633.564789] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 633.564789] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] self._vmops.spawn(context, instance, image_meta, injected_files, [ 633.564789] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 633.564789] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] vm_ref = self.build_virtual_machine(instance, [ 633.565249] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 633.565249] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] vif_infos = vmwarevif.get_vif_info(self._session, [ 633.565249] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 633.565249] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] for vif in network_info: [ 633.565249] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 633.565249] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] return self._sync_wrapper(fn, *args, **kwargs) [ 633.565249] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 633.565249] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] self.wait() [ 633.565249] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 633.565249] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] self[:] = self._gt.wait() [ 633.565249] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 633.565249] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] return self._exit_event.wait() [ 633.565249] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 633.565774] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] result = hub.switch() [ 633.565774] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 633.565774] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] return self.greenlet.switch() [ 633.565774] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.565774] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] result = function(*args, **kwargs) [ 633.565774] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 633.565774] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] return func(*args, **kwargs) [ 633.565774] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 633.565774] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] raise e [ 633.565774] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.565774] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] nwinfo = self.network_api.allocate_for_instance( [ 633.565774] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 633.565774] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] created_port_ids = self._update_ports_for_instance( [ 633.566781] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 633.566781] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] with excutils.save_and_reraise_exception(): [ 633.566781] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.566781] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] self.force_reraise() [ 633.566781] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.566781] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] raise self.value [ 633.566781] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 633.566781] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] updated_port = self._update_port( [ 633.566781] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.566781] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] _ensure_no_port_binding_failure(port) [ 633.566781] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.566781] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] raise exception.PortBindingFailed(port_id=port['id']) [ 633.567207] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] nova.exception.PortBindingFailed: Binding failed for port dc5d38ec-06fc-440f-bd69-0a197ea6da73, please check neutron logs for more information. [ 633.567207] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] [ 633.567207] env[62692]: INFO nova.compute.manager [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Terminating instance [ 633.567207] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Acquiring lock "refresh_cache-4fabba8d-25f5-48a0-a844-bdec6ea5c422" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.567207] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Acquired lock "refresh_cache-4fabba8d-25f5-48a0-a844-bdec6ea5c422" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.567207] env[62692]: DEBUG nova.network.neutron [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 633.608445] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.816089] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db24479-a737-4141-abe4-fa0cbb9c9704 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.824568] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39558589-22ae-40c0-a3e8-a456a0da7029 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.867034] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df42f00f-8892-4f6b-b626-0ef80d778543 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.874795] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af719d50-2c19-473f-8f81-791f42d580a7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.890551] env[62692]: DEBUG nova.compute.provider_tree [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 633.986624] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Acquiring lock "5d18922d-5937-47c8-a04a-aa757c153e8f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.986930] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Lock "5d18922d-5937-47c8-a04a-aa757c153e8f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.005545] env[62692]: DEBUG nova.compute.manager [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 634.072317] env[62692]: INFO nova.scheduler.client.report [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Deleted allocations for instance a16b5c50-a48a-4dfb-b2cb-bfe3723d108e [ 634.153306] env[62692]: DEBUG nova.network.neutron [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 634.394183] env[62692]: DEBUG nova.scheduler.client.report [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 634.589131] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e8fc4841-ad5f-4918-bbeb-39f7d2465d21 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "a16b5c50-a48a-4dfb-b2cb-bfe3723d108e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.232s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.590089] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "a16b5c50-a48a-4dfb-b2cb-bfe3723d108e" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 25.359s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.590796] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-62778ac9-6cfe-406f-a3c2-21b49a844863 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.606809] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f488d371-df9f-4301-92a7-56959a2b6cfe {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.626035] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "b76cce5f-b4ab-424f-b604-684f5c22154a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.626322] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "b76cce5f-b4ab-424f-b604-684f5c22154a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.899266] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.906s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.900399] env[62692]: ERROR nova.compute.manager [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b7e29e04-c6d2-4a03-bfdc-f406b818af64, please check neutron logs for more information. [ 634.900399] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Traceback (most recent call last): [ 634.900399] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 634.900399] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] self.driver.spawn(context, instance, image_meta, [ 634.900399] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 634.900399] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 634.900399] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 634.900399] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] vm_ref = self.build_virtual_machine(instance, [ 634.900399] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 634.900399] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] vif_infos = vmwarevif.get_vif_info(self._session, [ 634.900399] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 634.900796] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] for vif in network_info: [ 634.900796] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 634.900796] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] return self._sync_wrapper(fn, *args, **kwargs) [ 634.900796] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 634.900796] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] self.wait() [ 634.900796] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 634.900796] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] self[:] = self._gt.wait() [ 634.900796] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 634.900796] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] return self._exit_event.wait() [ 634.900796] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 634.900796] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] result = hub.switch() [ 634.900796] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 634.900796] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] return self.greenlet.switch() [ 634.901181] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.901181] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] result = function(*args, **kwargs) [ 634.901181] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 634.901181] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] return func(*args, **kwargs) [ 634.901181] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 634.901181] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] raise e [ 634.901181] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.901181] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] nwinfo = self.network_api.allocate_for_instance( [ 634.901181] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 634.901181] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] created_port_ids = self._update_ports_for_instance( [ 634.901181] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 634.901181] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] with excutils.save_and_reraise_exception(): [ 634.901181] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.901557] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] self.force_reraise() [ 634.901557] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.901557] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] raise self.value [ 634.901557] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 634.901557] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] updated_port = self._update_port( [ 634.901557] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.901557] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] _ensure_no_port_binding_failure(port) [ 634.901557] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.901557] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] raise exception.PortBindingFailed(port_id=port['id']) [ 634.901557] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] nova.exception.PortBindingFailed: Binding failed for port b7e29e04-c6d2-4a03-bfdc-f406b818af64, please check neutron logs for more information. [ 634.901557] env[62692]: ERROR nova.compute.manager [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] [ 634.904913] env[62692]: DEBUG nova.compute.utils [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Binding failed for port b7e29e04-c6d2-4a03-bfdc-f406b818af64, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 634.904913] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.106s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.909849] env[62692]: DEBUG nova.compute.manager [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Build of instance bf07e236-28b8-416e-94d3-fdf75cf62c5b was re-scheduled: Binding failed for port b7e29e04-c6d2-4a03-bfdc-f406b818af64, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 634.910566] env[62692]: DEBUG nova.compute.manager [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 634.910821] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Acquiring lock "refresh_cache-bf07e236-28b8-416e-94d3-fdf75cf62c5b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.910982] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Acquired lock "refresh_cache-bf07e236-28b8-416e-94d3-fdf75cf62c5b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.911922] env[62692]: DEBUG nova.network.neutron [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 634.919027] env[62692]: DEBUG nova.network.neutron [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.022048] env[62692]: DEBUG nova.compute.manager [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 635.069333] env[62692]: DEBUG nova.virt.hardware [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 635.069910] env[62692]: DEBUG nova.virt.hardware [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 635.070437] env[62692]: DEBUG nova.virt.hardware [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 635.070589] env[62692]: DEBUG nova.virt.hardware [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 635.071055] env[62692]: DEBUG nova.virt.hardware [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 635.073236] env[62692]: DEBUG nova.virt.hardware [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 635.073557] env[62692]: DEBUG nova.virt.hardware [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 635.073705] env[62692]: DEBUG nova.virt.hardware [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 635.073881] env[62692]: DEBUG nova.virt.hardware [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 635.074062] env[62692]: DEBUG nova.virt.hardware [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 635.074243] env[62692]: DEBUG nova.virt.hardware [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 635.075298] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9aa3b7d-8c00-4ceb-920f-c2b037cff593 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.088698] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d73f6324-98f3-4599-b9c2-8531257d7c8d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.098933] env[62692]: DEBUG nova.compute.manager [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 635.113353] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Instance VIF info [] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 635.131240] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 635.131655] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-beea23ba-4b59-4454-9c38-b75f6939318f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.141691] env[62692]: INFO nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: a16b5c50-a48a-4dfb-b2cb-bfe3723d108e] During the sync_power process the instance has moved from host None to host cpu-1 [ 635.141915] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "a16b5c50-a48a-4dfb-b2cb-bfe3723d108e" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.552s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.151148] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Created folder: OpenStack in parent group-v4. [ 635.151148] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Creating folder: Project (6ca368519abc4145aa9e3d03b3d94ce4). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 635.151148] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-de286e8c-5387-423c-b752-43b8516ed2b0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.164900] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Created folder: Project (6ca368519abc4145aa9e3d03b3d94ce4) in parent group-v248868. [ 635.169071] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Creating folder: Instances. Parent ref: group-v248869. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 635.169071] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8ae20bb2-cc8e-4c94-a81f-32ca6e762bff {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.182982] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Created folder: Instances in parent group-v248869. [ 635.183263] env[62692]: DEBUG oslo.service.loopingcall [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 635.183467] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 635.183669] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5ce83f37-7d28-4c0c-b0a6-c7ed392438db {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.213019] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 635.213019] env[62692]: value = "task-1140873" [ 635.213019] env[62692]: _type = "Task" [ 635.213019] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.226516] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140873, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.258109] env[62692]: DEBUG nova.compute.manager [req-2ac32085-356e-4964-b485-b97c038bc2d4 req-047e4699-534d-4872-b5d4-5049312ae26c service nova] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Received event network-changed-dc5d38ec-06fc-440f-bd69-0a197ea6da73 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 635.258331] env[62692]: DEBUG nova.compute.manager [req-2ac32085-356e-4964-b485-b97c038bc2d4 req-047e4699-534d-4872-b5d4-5049312ae26c service nova] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Refreshing instance network info cache due to event network-changed-dc5d38ec-06fc-440f-bd69-0a197ea6da73. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 635.258522] env[62692]: DEBUG oslo_concurrency.lockutils [req-2ac32085-356e-4964-b485-b97c038bc2d4 req-047e4699-534d-4872-b5d4-5049312ae26c service nova] Acquiring lock "refresh_cache-4fabba8d-25f5-48a0-a844-bdec6ea5c422" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.424984] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Releasing lock "refresh_cache-4fabba8d-25f5-48a0-a844-bdec6ea5c422" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.425287] env[62692]: DEBUG nova.compute.manager [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 635.425367] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 635.425655] env[62692]: DEBUG oslo_concurrency.lockutils [req-2ac32085-356e-4964-b485-b97c038bc2d4 req-047e4699-534d-4872-b5d4-5049312ae26c service nova] Acquired lock "refresh_cache-4fabba8d-25f5-48a0-a844-bdec6ea5c422" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.425865] env[62692]: DEBUG nova.network.neutron [req-2ac32085-356e-4964-b485-b97c038bc2d4 req-047e4699-534d-4872-b5d4-5049312ae26c service nova] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Refreshing network info cache for port dc5d38ec-06fc-440f-bd69-0a197ea6da73 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 635.429399] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1ccc13e6-1465-4225-9913-b82635683c6e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.443034] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5d4a0c7-b3ae-4991-90b2-c46dbfb96de0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.468222] env[62692]: DEBUG nova.network.neutron [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.490249] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4fabba8d-25f5-48a0-a844-bdec6ea5c422 could not be found. [ 635.490472] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 635.490751] env[62692]: INFO nova.compute.manager [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Took 0.07 seconds to destroy the instance on the hypervisor. [ 635.491220] env[62692]: DEBUG oslo.service.loopingcall [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 635.494578] env[62692]: DEBUG nova.compute.manager [-] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 635.494760] env[62692]: DEBUG nova.network.neutron [-] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 635.572043] env[62692]: DEBUG nova.network.neutron [-] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.637176] env[62692]: DEBUG oslo_concurrency.lockutils [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.730923] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140873, 'name': CreateVM_Task, 'duration_secs': 0.42841} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.730923] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 635.731235] env[62692]: DEBUG oslo_vmware.service [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33511184-2993-4d04-9402-ae7627e6a957 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.738452] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.738630] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.739349] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 635.739619] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91ec67c7-0eb6-4021-9151-a6703434b9f8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.750741] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Waiting for the task: (returnval){ [ 635.750741] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]521d9ba5-7723-b98f-dba8-6dcc2b43aeb2" [ 635.750741] env[62692]: _type = "Task" [ 635.750741] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.759263] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]521d9ba5-7723-b98f-dba8-6dcc2b43aeb2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.804499] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23f5890f-4003-4ea6-ae2d-59c1bd6c84d7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.812821] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86707258-f714-4591-b45e-bb084e94ec41 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.848207] env[62692]: DEBUG nova.network.neutron [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.849409] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16f1c723-e112-42a2-a69d-6cece02ec0d5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.858285] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e04ae2cc-b3f4-4e63-8de7-e43cada3caf7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.876657] env[62692]: DEBUG nova.compute.provider_tree [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 635.980813] env[62692]: DEBUG nova.network.neutron [req-2ac32085-356e-4964-b485-b97c038bc2d4 req-047e4699-534d-4872-b5d4-5049312ae26c service nova] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.076167] env[62692]: DEBUG nova.network.neutron [-] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.262055] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.262330] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 636.262566] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.262712] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.263138] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 636.263399] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8224269f-5094-4d5f-9d8d-cab840867329 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.280829] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 636.281090] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 636.282150] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-414de727-2c06-448f-b3fb-4016862a87b9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.289967] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5ce0e7b-68c5-4043-a421-a35e3ab5b12a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.295791] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Waiting for the task: (returnval){ [ 636.295791] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]521ed605-173c-4155-67a0-d4c062e30bba" [ 636.295791] env[62692]: _type = "Task" [ 636.295791] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.303324] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]521ed605-173c-4155-67a0-d4c062e30bba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.325091] env[62692]: DEBUG nova.network.neutron [req-2ac32085-356e-4964-b485-b97c038bc2d4 req-047e4699-534d-4872-b5d4-5049312ae26c service nova] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.356227] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Releasing lock "refresh_cache-bf07e236-28b8-416e-94d3-fdf75cf62c5b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.356227] env[62692]: DEBUG nova.compute.manager [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 636.356227] env[62692]: DEBUG nova.compute.manager [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 636.358141] env[62692]: DEBUG nova.network.neutron [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 636.383088] env[62692]: DEBUG nova.scheduler.client.report [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 636.389785] env[62692]: DEBUG nova.network.neutron [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.580819] env[62692]: INFO nova.compute.manager [-] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Took 1.08 seconds to deallocate network for instance. [ 636.586729] env[62692]: DEBUG nova.compute.claims [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 636.586948] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.810257] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Preparing fetch location {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 636.810257] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Creating directory with path [datastore2] vmware_temp/40e70add-5223-456a-8f95-a2d71d180a17/81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 636.810257] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-02ac01da-c4db-47ad-8570-34d936ed4a53 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.827671] env[62692]: DEBUG oslo_concurrency.lockutils [req-2ac32085-356e-4964-b485-b97c038bc2d4 req-047e4699-534d-4872-b5d4-5049312ae26c service nova] Releasing lock "refresh_cache-4fabba8d-25f5-48a0-a844-bdec6ea5c422" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.844520] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Created directory with path [datastore2] vmware_temp/40e70add-5223-456a-8f95-a2d71d180a17/81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 636.844857] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Fetch image to [datastore2] vmware_temp/40e70add-5223-456a-8f95-a2d71d180a17/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/tmp-sparse.vmdk {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 636.845208] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Downloading image file data 81eb64fb-5a0c-47b7-9948-7dcb01b911bd to [datastore2] vmware_temp/40e70add-5223-456a-8f95-a2d71d180a17/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/tmp-sparse.vmdk on the data store datastore2 {{(pid=62692) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 636.846082] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1448dac-ce1e-4f35-9c04-6702343d5d13 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.855772] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-531da44a-cdcc-4bee-973a-cd9640a717b1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.867927] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b3728b8-794e-4fb7-aa1e-95f4fa531d22 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.908875] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.004s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.909561] env[62692]: ERROR nova.compute.manager [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b344a4c1-5655-4faf-8f81-fa793eaf8eb6, please check neutron logs for more information. [ 636.909561] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Traceback (most recent call last): [ 636.909561] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 636.909561] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] self.driver.spawn(context, instance, image_meta, [ 636.909561] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 636.909561] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 636.909561] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 636.909561] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] vm_ref = self.build_virtual_machine(instance, [ 636.909561] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 636.909561] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] vif_infos = vmwarevif.get_vif_info(self._session, [ 636.909561] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 636.911135] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] for vif in network_info: [ 636.911135] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 636.911135] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] return self._sync_wrapper(fn, *args, **kwargs) [ 636.911135] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 636.911135] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] self.wait() [ 636.911135] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 636.911135] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] self[:] = self._gt.wait() [ 636.911135] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 636.911135] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] return self._exit_event.wait() [ 636.911135] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 636.911135] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] result = hub.switch() [ 636.911135] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 636.911135] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] return self.greenlet.switch() [ 636.911569] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 636.911569] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] result = function(*args, **kwargs) [ 636.911569] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 636.911569] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] return func(*args, **kwargs) [ 636.911569] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 636.911569] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] raise e [ 636.911569] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 636.911569] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] nwinfo = self.network_api.allocate_for_instance( [ 636.911569] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 636.911569] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] created_port_ids = self._update_ports_for_instance( [ 636.911569] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 636.911569] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] with excutils.save_and_reraise_exception(): [ 636.911569] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.911904] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] self.force_reraise() [ 636.911904] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.911904] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] raise self.value [ 636.911904] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 636.911904] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] updated_port = self._update_port( [ 636.911904] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.911904] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] _ensure_no_port_binding_failure(port) [ 636.911904] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.911904] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] raise exception.PortBindingFailed(port_id=port['id']) [ 636.911904] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] nova.exception.PortBindingFailed: Binding failed for port b344a4c1-5655-4faf-8f81-fa793eaf8eb6, please check neutron logs for more information. [ 636.911904] env[62692]: ERROR nova.compute.manager [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] [ 636.912208] env[62692]: DEBUG nova.compute.utils [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Binding failed for port b344a4c1-5655-4faf-8f81-fa793eaf8eb6, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 636.912208] env[62692]: DEBUG nova.network.neutron [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.913352] env[62692]: DEBUG oslo_concurrency.lockutils [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.228s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.916751] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63a289b7-b316-4cd0-a8e2-8b4b857e30cb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.920305] env[62692]: DEBUG nova.compute.manager [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Build of instance 2d5a412a-66c0-46fc-84bc-2edb90f4593c was re-scheduled: Binding failed for port b344a4c1-5655-4faf-8f81-fa793eaf8eb6, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 636.920768] env[62692]: DEBUG nova.compute.manager [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 636.921790] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Acquiring lock "refresh_cache-2d5a412a-66c0-46fc-84bc-2edb90f4593c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.921790] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Acquired lock "refresh_cache-2d5a412a-66c0-46fc-84bc-2edb90f4593c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.921790] env[62692]: DEBUG nova.network.neutron [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 636.926900] env[62692]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-0bca8bb2-608c-4846-aa40-b49c3cf50a5c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.015479] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Downloading image file data 81eb64fb-5a0c-47b7-9948-7dcb01b911bd to the data store datastore2 {{(pid=62692) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 637.094128] env[62692]: DEBUG oslo_vmware.rw_handles [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/40e70add-5223-456a-8f95-a2d71d180a17/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62692) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 637.429873] env[62692]: INFO nova.compute.manager [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] [instance: bf07e236-28b8-416e-94d3-fdf75cf62c5b] Took 1.07 seconds to deallocate network for instance. [ 637.513214] env[62692]: DEBUG nova.network.neutron [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 637.741811] env[62692]: DEBUG nova.network.neutron [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.833359] env[62692]: DEBUG oslo_vmware.rw_handles [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Completed reading data from the image iterator. {{(pid=62692) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 637.834030] env[62692]: DEBUG oslo_vmware.rw_handles [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Closing write handle for https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/40e70add-5223-456a-8f95-a2d71d180a17/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62692) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 637.896417] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50f54f96-0c82-48c4-ae26-aaff9ea35048 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.905880] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afa4be3a-124f-4721-8744-26f8b86d6282 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.940956] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c84003b5-29d2-4daa-a740-ba2e7f50f04b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.954709] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39c2b46a-e709-4bcc-ae3d-844ddcb5e388 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.974211] env[62692]: DEBUG nova.compute.provider_tree [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.976453] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Downloaded image file data 81eb64fb-5a0c-47b7-9948-7dcb01b911bd to vmware_temp/40e70add-5223-456a-8f95-a2d71d180a17/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/tmp-sparse.vmdk on the data store datastore2 {{(pid=62692) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 637.980855] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Caching image {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 637.980855] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Copying Virtual Disk [datastore2] vmware_temp/40e70add-5223-456a-8f95-a2d71d180a17/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/tmp-sparse.vmdk to [datastore2] vmware_temp/40e70add-5223-456a-8f95-a2d71d180a17/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 637.980855] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-12191d26-29af-4c8c-b175-1131c6095397 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.992126] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Waiting for the task: (returnval){ [ 637.992126] env[62692]: value = "task-1140874" [ 637.992126] env[62692]: _type = "Task" [ 637.992126] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.005414] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': task-1140874, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.246940] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Releasing lock "refresh_cache-2d5a412a-66c0-46fc-84bc-2edb90f4593c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.247441] env[62692]: DEBUG nova.compute.manager [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 638.247960] env[62692]: DEBUG nova.compute.manager [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 638.248392] env[62692]: DEBUG nova.network.neutron [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 638.318731] env[62692]: DEBUG nova.network.neutron [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 638.487115] env[62692]: DEBUG nova.scheduler.client.report [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 638.506759] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': task-1140874, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.511408] env[62692]: INFO nova.scheduler.client.report [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Deleted allocations for instance bf07e236-28b8-416e-94d3-fdf75cf62c5b [ 638.824230] env[62692]: DEBUG nova.network.neutron [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.994633] env[62692]: DEBUG oslo_concurrency.lockutils [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.082s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.995372] env[62692]: ERROR nova.compute.manager [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 466bf122-912c-490c-9871-bb2062bf6b3c, please check neutron logs for more information. [ 638.995372] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Traceback (most recent call last): [ 638.995372] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 638.995372] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] self.driver.spawn(context, instance, image_meta, [ 638.995372] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 638.995372] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 638.995372] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 638.995372] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] vm_ref = self.build_virtual_machine(instance, [ 638.995372] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 638.995372] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] vif_infos = vmwarevif.get_vif_info(self._session, [ 638.995372] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 638.996273] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] for vif in network_info: [ 638.996273] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 638.996273] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] return self._sync_wrapper(fn, *args, **kwargs) [ 638.996273] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 638.996273] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] self.wait() [ 638.996273] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 638.996273] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] self[:] = self._gt.wait() [ 638.996273] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 638.996273] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] return self._exit_event.wait() [ 638.996273] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 638.996273] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] result = hub.switch() [ 638.996273] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 638.996273] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] return self.greenlet.switch() [ 638.997496] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.997496] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] result = function(*args, **kwargs) [ 638.997496] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 638.997496] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] return func(*args, **kwargs) [ 638.997496] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 638.997496] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] raise e [ 638.997496] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.997496] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] nwinfo = self.network_api.allocate_for_instance( [ 638.997496] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 638.997496] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] created_port_ids = self._update_ports_for_instance( [ 638.997496] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 638.997496] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] with excutils.save_and_reraise_exception(): [ 638.997496] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.997996] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] self.force_reraise() [ 638.997996] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.997996] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] raise self.value [ 638.997996] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 638.997996] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] updated_port = self._update_port( [ 638.997996] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.997996] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] _ensure_no_port_binding_failure(port) [ 638.997996] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.997996] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] raise exception.PortBindingFailed(port_id=port['id']) [ 638.997996] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] nova.exception.PortBindingFailed: Binding failed for port 466bf122-912c-490c-9871-bb2062bf6b3c, please check neutron logs for more information. [ 638.997996] env[62692]: ERROR nova.compute.manager [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] [ 638.998318] env[62692]: DEBUG nova.compute.utils [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Binding failed for port 466bf122-912c-490c-9871-bb2062bf6b3c, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 639.002123] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.161s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.002123] env[62692]: INFO nova.compute.claims [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 639.017033] env[62692]: DEBUG nova.compute.manager [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Build of instance 331411c2-1758-46a4-ae08-8c575d0b8be2 was re-scheduled: Binding failed for port 466bf122-912c-490c-9871-bb2062bf6b3c, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 639.017033] env[62692]: DEBUG nova.compute.manager [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 639.017033] env[62692]: DEBUG oslo_concurrency.lockutils [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Acquiring lock "refresh_cache-331411c2-1758-46a4-ae08-8c575d0b8be2" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.017033] env[62692]: DEBUG oslo_concurrency.lockutils [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Acquired lock "refresh_cache-331411c2-1758-46a4-ae08-8c575d0b8be2" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.017276] env[62692]: DEBUG nova.network.neutron [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 639.022387] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09f01855-6d81-4ddf-b580-915c63b3d36e tempest-AttachInterfacesV270Test-1019685148 tempest-AttachInterfacesV270Test-1019685148-project-member] Lock "bf07e236-28b8-416e-94d3-fdf75cf62c5b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.526s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.033273] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': task-1140874, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.707505} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.034343] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Copied Virtual Disk [datastore2] vmware_temp/40e70add-5223-456a-8f95-a2d71d180a17/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/tmp-sparse.vmdk to [datastore2] vmware_temp/40e70add-5223-456a-8f95-a2d71d180a17/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 639.035732] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Deleting the datastore file [datastore2] vmware_temp/40e70add-5223-456a-8f95-a2d71d180a17/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/tmp-sparse.vmdk {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 639.036872] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-69fd4915-81d6-459f-85cb-6e58f60f3bf0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.054323] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Waiting for the task: (returnval){ [ 639.054323] env[62692]: value = "task-1140875" [ 639.054323] env[62692]: _type = "Task" [ 639.054323] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.062258] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': task-1140875, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.071702] env[62692]: DEBUG nova.compute.manager [req-f50e385f-d191-432c-83cb-ba668672f4fd req-697e1789-23ea-474f-9b05-3bc1cbbc82be service nova] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Received event network-vif-deleted-dc5d38ec-06fc-440f-bd69-0a197ea6da73 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 639.329259] env[62692]: INFO nova.compute.manager [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] [instance: 2d5a412a-66c0-46fc-84bc-2edb90f4593c] Took 1.08 seconds to deallocate network for instance. [ 639.527787] env[62692]: DEBUG nova.compute.manager [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 639.568353] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': task-1140875, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024455} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.570253] env[62692]: DEBUG nova.network.neutron [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 639.573540] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 639.573540] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Moving file from [datastore2] vmware_temp/40e70add-5223-456a-8f95-a2d71d180a17/81eb64fb-5a0c-47b7-9948-7dcb01b911bd to [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd. {{(pid=62692) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 639.573540] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-86d5bb54-b803-4a00-973e-c04f004bb494 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.583073] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Waiting for the task: (returnval){ [ 639.583073] env[62692]: value = "task-1140876" [ 639.583073] env[62692]: _type = "Task" [ 639.583073] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.594456] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': task-1140876, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.718240] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Acquiring lock "2feffd08-f011-4117-9f8d-ac6730f893f5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.718240] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Lock "2feffd08-f011-4117-9f8d-ac6730f893f5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.718240] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Acquiring lock "9422a9a8-795d-4794-8ba9-9e6f88cdb464" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.718240] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Lock "9422a9a8-795d-4794-8ba9-9e6f88cdb464" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.813309] env[62692]: DEBUG nova.network.neutron [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.070988] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.099093] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': task-1140876, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.026067} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.104803] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] File moved {{(pid=62692) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 640.105073] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Cleaning up location [datastore2] vmware_temp/40e70add-5223-456a-8f95-a2d71d180a17 {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 640.105937] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Deleting the datastore file [datastore2] vmware_temp/40e70add-5223-456a-8f95-a2d71d180a17 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 640.105937] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4cbd1a48-88de-4065-8442-1193260a6b29 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.119473] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Waiting for the task: (returnval){ [ 640.119473] env[62692]: value = "task-1140877" [ 640.119473] env[62692]: _type = "Task" [ 640.119473] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.141290] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': task-1140877, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.299326] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquiring lock "146e164e-2543-46ab-a7d3-9803d56f61ec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.299326] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Lock "146e164e-2543-46ab-a7d3-9803d56f61ec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.316893] env[62692]: DEBUG oslo_concurrency.lockutils [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Releasing lock "refresh_cache-331411c2-1758-46a4-ae08-8c575d0b8be2" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.319142] env[62692]: DEBUG nova.compute.manager [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 640.319142] env[62692]: DEBUG nova.compute.manager [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 640.319142] env[62692]: DEBUG nova.network.neutron [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 640.366855] env[62692]: DEBUG nova.network.neutron [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 640.376693] env[62692]: INFO nova.scheduler.client.report [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Deleted allocations for instance 2d5a412a-66c0-46fc-84bc-2edb90f4593c [ 640.458748] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73125fb2-bb4f-476c-9e23-7113a00ea4b0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.471691] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c871e33-612d-4599-882c-56a6768a060a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.513076] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92b06d72-2278-4e3b-a9b7-c2aac256a3ee {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.522423] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc78425c-360c-449f-82c9-f6365c924e0c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.538503] env[62692]: DEBUG nova.compute.provider_tree [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 640.633034] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': task-1140877, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.038748} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.633034] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 640.633034] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7d6ee8d-6838-4682-9394-009fc70bb83b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.640051] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Waiting for the task: (returnval){ [ 640.640051] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5247eb31-9945-741d-de37-0ee14779f6bc" [ 640.640051] env[62692]: _type = "Task" [ 640.640051] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.648611] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5247eb31-9945-741d-de37-0ee14779f6bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.873480] env[62692]: DEBUG nova.network.neutron [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.889463] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1a892fa9-c196-4044-81da-a84ec2f43cf7 tempest-ServerDiagnosticsTest-1226470871 tempest-ServerDiagnosticsTest-1226470871-project-member] Lock "2d5a412a-66c0-46fc-84bc-2edb90f4593c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.375s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.046794] env[62692]: DEBUG nova.scheduler.client.report [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 641.151814] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5247eb31-9945-741d-de37-0ee14779f6bc, 'name': SearchDatastore_Task, 'duration_secs': 0.012342} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.153146] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.153411] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 54f615e3-5c6e-44bb-aeb7-5620a639b55c/54f615e3-5c6e-44bb-aeb7-5620a639b55c.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 641.153684] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4e437fb4-8c32-4132-87bc-281a6643b73c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.168035] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Waiting for the task: (returnval){ [ 641.168035] env[62692]: value = "task-1140878" [ 641.168035] env[62692]: _type = "Task" [ 641.168035] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.176185] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': task-1140878, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.377725] env[62692]: INFO nova.compute.manager [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] [instance: 331411c2-1758-46a4-ae08-8c575d0b8be2] Took 1.06 seconds to deallocate network for instance. [ 641.392351] env[62692]: DEBUG nova.compute.manager [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 641.552746] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.553s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.552849] env[62692]: DEBUG nova.compute.manager [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 641.556557] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.890s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.680910] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': task-1140878, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.925234] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.034443] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "dbd51115-6b41-4019-955d-16a52b534e84" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.034917] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "dbd51115-6b41-4019-955d-16a52b534e84" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.060021] env[62692]: DEBUG nova.compute.utils [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 642.060021] env[62692]: DEBUG nova.compute.manager [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 642.060021] env[62692]: DEBUG nova.network.neutron [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 642.181647] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': task-1140878, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.662496} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.183228] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 54f615e3-5c6e-44bb-aeb7-5620a639b55c/54f615e3-5c6e-44bb-aeb7-5620a639b55c.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 642.183479] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 642.186086] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f2ee2e86-a9f5-4f2b-a2d1-b383a90cdf86 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.194040] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Waiting for the task: (returnval){ [ 642.194040] env[62692]: value = "task-1140879" [ 642.194040] env[62692]: _type = "Task" [ 642.194040] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.203995] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': task-1140879, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.373122] env[62692]: DEBUG nova.policy [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8adae7ddecbc4aa28694a228f52b5bd6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b00f181776484ecebc81d7469d81153f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 642.421835] env[62692]: INFO nova.scheduler.client.report [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Deleted allocations for instance 331411c2-1758-46a4-ae08-8c575d0b8be2 [ 642.431092] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-831d2e18-ad41-4cbe-b36c-013a413933ec {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.438816] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dcab1e1-8f4c-4fd8-9ef0-c78f00665387 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.484705] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3ad5498-02ba-47c3-a765-7982ac5f7d4d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.494064] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d71f7c4e-7b27-46a8-a902-a4246b189be5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.510747] env[62692]: DEBUG nova.compute.provider_tree [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 642.564779] env[62692]: DEBUG nova.compute.manager [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 642.712147] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': task-1140879, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063113} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.714981] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 642.716064] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-696508f0-acda-48a7-956b-254a5c3ba59e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.742885] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Reconfiguring VM instance instance-0000000a to attach disk [datastore2] 54f615e3-5c6e-44bb-aeb7-5620a639b55c/54f615e3-5c6e-44bb-aeb7-5620a639b55c.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 642.742885] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-98be74e5-bee0-43c9-a8a7-136116d8e9c8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.770321] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Waiting for the task: (returnval){ [ 642.770321] env[62692]: value = "task-1140880" [ 642.770321] env[62692]: _type = "Task" [ 642.770321] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.774478] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': task-1140880, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.934405] env[62692]: DEBUG oslo_concurrency.lockutils [None req-514fd557-cdaf-4edb-ba91-4f39c7ccc8e3 tempest-ServersAdminNegativeTestJSON-1752473742 tempest-ServersAdminNegativeTestJSON-1752473742-project-member] Lock "331411c2-1758-46a4-ae08-8c575d0b8be2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.300s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 643.016437] env[62692]: DEBUG nova.scheduler.client.report [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 643.275660] env[62692]: DEBUG nova.network.neutron [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Successfully created port: d37419d2-13ae-4fd0-8880-7dd92681d221 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 643.280100] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': task-1140880, 'name': ReconfigVM_Task, 'duration_secs': 0.310472} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.280674] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Reconfigured VM instance instance-0000000a to attach disk [datastore2] 54f615e3-5c6e-44bb-aeb7-5620a639b55c/54f615e3-5c6e-44bb-aeb7-5620a639b55c.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 643.281849] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0d9cf76b-b6fa-4678-93ba-1f0dff0121e4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.290097] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Waiting for the task: (returnval){ [ 643.290097] env[62692]: value = "task-1140881" [ 643.290097] env[62692]: _type = "Task" [ 643.290097] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.297309] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': task-1140881, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.440875] env[62692]: DEBUG nova.compute.manager [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 643.520420] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.964s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 643.521189] env[62692]: ERROR nova.compute.manager [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1d2e6b57-3cfb-4d93-95e8-f855731915ed, please check neutron logs for more information. [ 643.521189] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Traceback (most recent call last): [ 643.521189] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 643.521189] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] self.driver.spawn(context, instance, image_meta, [ 643.521189] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 643.521189] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 643.521189] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 643.521189] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] vm_ref = self.build_virtual_machine(instance, [ 643.521189] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 643.521189] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] vif_infos = vmwarevif.get_vif_info(self._session, [ 643.521189] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 643.521555] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] for vif in network_info: [ 643.521555] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 643.521555] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] return self._sync_wrapper(fn, *args, **kwargs) [ 643.521555] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 643.521555] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] self.wait() [ 643.521555] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 643.521555] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] self[:] = self._gt.wait() [ 643.521555] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 643.521555] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] return self._exit_event.wait() [ 643.521555] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 643.521555] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] result = hub.switch() [ 643.521555] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 643.521555] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] return self.greenlet.switch() [ 643.521916] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 643.521916] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] result = function(*args, **kwargs) [ 643.521916] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 643.521916] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] return func(*args, **kwargs) [ 643.521916] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 643.521916] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] raise e [ 643.521916] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 643.521916] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] nwinfo = self.network_api.allocate_for_instance( [ 643.521916] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 643.521916] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] created_port_ids = self._update_ports_for_instance( [ 643.521916] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 643.521916] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] with excutils.save_and_reraise_exception(): [ 643.521916] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 643.522290] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] self.force_reraise() [ 643.522290] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 643.522290] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] raise self.value [ 643.522290] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 643.522290] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] updated_port = self._update_port( [ 643.522290] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 643.522290] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] _ensure_no_port_binding_failure(port) [ 643.522290] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 643.522290] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] raise exception.PortBindingFailed(port_id=port['id']) [ 643.522290] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] nova.exception.PortBindingFailed: Binding failed for port 1d2e6b57-3cfb-4d93-95e8-f855731915ed, please check neutron logs for more information. [ 643.522290] env[62692]: ERROR nova.compute.manager [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] [ 643.522611] env[62692]: DEBUG nova.compute.utils [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Binding failed for port 1d2e6b57-3cfb-4d93-95e8-f855731915ed, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 643.524827] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.226s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.529581] env[62692]: DEBUG nova.compute.manager [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Build of instance 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e was re-scheduled: Binding failed for port 1d2e6b57-3cfb-4d93-95e8-f855731915ed, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 643.529728] env[62692]: DEBUG nova.compute.manager [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 643.530250] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Acquiring lock "refresh_cache-5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 643.530250] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Acquired lock "refresh_cache-5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.530250] env[62692]: DEBUG nova.network.neutron [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 643.574640] env[62692]: DEBUG nova.compute.manager [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 643.606731] env[62692]: DEBUG nova.virt.hardware [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 643.607288] env[62692]: DEBUG nova.virt.hardware [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 643.607288] env[62692]: DEBUG nova.virt.hardware [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 643.607420] env[62692]: DEBUG nova.virt.hardware [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 643.607496] env[62692]: DEBUG nova.virt.hardware [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 643.607642] env[62692]: DEBUG nova.virt.hardware [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 643.607975] env[62692]: DEBUG nova.virt.hardware [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 643.608101] env[62692]: DEBUG nova.virt.hardware [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 643.608338] env[62692]: DEBUG nova.virt.hardware [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 643.608558] env[62692]: DEBUG nova.virt.hardware [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 643.608778] env[62692]: DEBUG nova.virt.hardware [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 643.609787] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e6660ac-a0c2-46f0-b64a-9b9c0b80c98e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.620923] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb216006-e7dc-4054-99c8-5a5690150d5b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.680742] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Acquiring lock "49d69437-7046-4cb2-914c-8544dca70d3f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.681237] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Lock "49d69437-7046-4cb2-914c-8544dca70d3f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.721448] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Acquiring lock "db727822-af78-49bc-98d9-9caabc35b3b6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.721448] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Lock "db727822-af78-49bc-98d9-9caabc35b3b6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.746629] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Acquiring lock "0266b67a-7c53-471d-a167-ac7362c4b9be" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.746874] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Lock "0266b67a-7c53-471d-a167-ac7362c4b9be" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.802239] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': task-1140881, 'name': Rename_Task, 'duration_secs': 0.132935} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.802239] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 643.802239] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c4e3e076-15f6-4678-a6f7-8f9eecf2ea79 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.809067] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Waiting for the task: (returnval){ [ 643.809067] env[62692]: value = "task-1140882" [ 643.809067] env[62692]: _type = "Task" [ 643.809067] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.817522] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': task-1140882, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.973303] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.100863] env[62692]: DEBUG nova.network.neutron [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 644.320661] env[62692]: DEBUG oslo_vmware.api [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': task-1140882, 'name': PowerOnVM_Task, 'duration_secs': 0.417338} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.323922] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 644.323922] env[62692]: INFO nova.compute.manager [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Took 9.30 seconds to spawn the instance on the hypervisor. [ 644.324151] env[62692]: DEBUG nova.compute.manager [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 644.325016] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e28cb21-95a2-4a9f-aaf9-bd71c2de3443 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.359989] env[62692]: DEBUG nova.network.neutron [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.399666] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e77e7ca-0354-410c-b852-a97182e39715 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.407320] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c25f6609-88b1-426e-9ba5-c3c633134f4c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.449142] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31c742ad-3751-4ca1-8dc9-ab96f6c8ce8d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.458249] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-264357e3-a893-4944-a8f7-9273dd63736f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.472471] env[62692]: DEBUG nova.compute.provider_tree [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 644.851450] env[62692]: INFO nova.compute.manager [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Took 29.43 seconds to build instance. [ 644.862948] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Releasing lock "refresh_cache-5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.863219] env[62692]: DEBUG nova.compute.manager [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 644.863405] env[62692]: DEBUG nova.compute.manager [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 644.863570] env[62692]: DEBUG nova.network.neutron [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 644.957654] env[62692]: ERROR nova.compute.manager [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d37419d2-13ae-4fd0-8880-7dd92681d221, please check neutron logs for more information. [ 644.957654] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 644.957654] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.957654] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 644.957654] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 644.957654] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 644.957654] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 644.957654] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 644.957654] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.957654] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 644.957654] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.957654] env[62692]: ERROR nova.compute.manager raise self.value [ 644.957654] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 644.957654] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 644.957654] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.957654] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 644.958342] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.958342] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 644.958342] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d37419d2-13ae-4fd0-8880-7dd92681d221, please check neutron logs for more information. [ 644.958342] env[62692]: ERROR nova.compute.manager [ 644.958342] env[62692]: Traceback (most recent call last): [ 644.958342] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 644.958342] env[62692]: listener.cb(fileno) [ 644.958342] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 644.958342] env[62692]: result = function(*args, **kwargs) [ 644.958342] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 644.958342] env[62692]: return func(*args, **kwargs) [ 644.958342] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 644.958342] env[62692]: raise e [ 644.958342] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.958342] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 644.958342] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 644.958342] env[62692]: created_port_ids = self._update_ports_for_instance( [ 644.958342] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 644.958342] env[62692]: with excutils.save_and_reraise_exception(): [ 644.958342] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.958342] env[62692]: self.force_reraise() [ 644.958342] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.958342] env[62692]: raise self.value [ 644.958342] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 644.958342] env[62692]: updated_port = self._update_port( [ 644.958342] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.958342] env[62692]: _ensure_no_port_binding_failure(port) [ 644.958342] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.958342] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 644.959826] env[62692]: nova.exception.PortBindingFailed: Binding failed for port d37419d2-13ae-4fd0-8880-7dd92681d221, please check neutron logs for more information. [ 644.959826] env[62692]: Removing descriptor: 19 [ 644.959826] env[62692]: ERROR nova.compute.manager [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d37419d2-13ae-4fd0-8880-7dd92681d221, please check neutron logs for more information. [ 644.959826] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Traceback (most recent call last): [ 644.959826] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 644.959826] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] yield resources [ 644.959826] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 644.959826] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] self.driver.spawn(context, instance, image_meta, [ 644.959826] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 644.959826] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] self._vmops.spawn(context, instance, image_meta, injected_files, [ 644.959826] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 644.959826] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] vm_ref = self.build_virtual_machine(instance, [ 644.960401] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 644.960401] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] vif_infos = vmwarevif.get_vif_info(self._session, [ 644.960401] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 644.960401] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] for vif in network_info: [ 644.960401] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 644.960401] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] return self._sync_wrapper(fn, *args, **kwargs) [ 644.960401] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 644.960401] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] self.wait() [ 644.960401] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 644.960401] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] self[:] = self._gt.wait() [ 644.960401] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 644.960401] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] return self._exit_event.wait() [ 644.960401] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 644.960983] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] result = hub.switch() [ 644.960983] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 644.960983] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] return self.greenlet.switch() [ 644.960983] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 644.960983] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] result = function(*args, **kwargs) [ 644.960983] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 644.960983] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] return func(*args, **kwargs) [ 644.960983] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 644.960983] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] raise e [ 644.960983] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.960983] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] nwinfo = self.network_api.allocate_for_instance( [ 644.960983] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 644.960983] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] created_port_ids = self._update_ports_for_instance( [ 644.961437] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 644.961437] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] with excutils.save_and_reraise_exception(): [ 644.961437] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.961437] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] self.force_reraise() [ 644.961437] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.961437] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] raise self.value [ 644.961437] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 644.961437] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] updated_port = self._update_port( [ 644.961437] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.961437] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] _ensure_no_port_binding_failure(port) [ 644.961437] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.961437] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] raise exception.PortBindingFailed(port_id=port['id']) [ 644.961894] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] nova.exception.PortBindingFailed: Binding failed for port d37419d2-13ae-4fd0-8880-7dd92681d221, please check neutron logs for more information. [ 644.961894] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] [ 644.961894] env[62692]: INFO nova.compute.manager [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Terminating instance [ 644.964671] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Acquiring lock "refresh_cache-f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 644.964999] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Acquired lock "refresh_cache-f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.965162] env[62692]: DEBUG nova.network.neutron [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 644.974897] env[62692]: DEBUG nova.scheduler.client.report [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 645.026197] env[62692]: DEBUG nova.network.neutron [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 645.062035] env[62692]: DEBUG oslo_concurrency.lockutils [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Acquiring lock "a5fa20ab-66b7-4632-8a80-ac031263c042" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.062270] env[62692]: DEBUG oslo_concurrency.lockutils [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Lock "a5fa20ab-66b7-4632-8a80-ac031263c042" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.359894] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1d846b2-b4b0-42ce-a910-2a30bf8a6c1d tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Lock "54f615e3-5c6e-44bb-aeb7-5620a639b55c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.953s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.479346] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.954s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.480625] env[62692]: ERROR nova.compute.manager [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d33d9daf-8548-4f1d-a09e-ed9348cf5d0c, please check neutron logs for more information. [ 645.480625] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] Traceback (most recent call last): [ 645.480625] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 645.480625] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] self.driver.spawn(context, instance, image_meta, [ 645.480625] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 645.480625] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] self._vmops.spawn(context, instance, image_meta, injected_files, [ 645.480625] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 645.480625] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] vm_ref = self.build_virtual_machine(instance, [ 645.480625] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 645.480625] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] vif_infos = vmwarevif.get_vif_info(self._session, [ 645.480625] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 645.481232] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] for vif in network_info: [ 645.481232] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 645.481232] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] return self._sync_wrapper(fn, *args, **kwargs) [ 645.481232] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 645.481232] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] self.wait() [ 645.481232] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 645.481232] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] self[:] = self._gt.wait() [ 645.481232] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 645.481232] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] return self._exit_event.wait() [ 645.481232] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 645.481232] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] result = hub.switch() [ 645.481232] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 645.481232] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] return self.greenlet.switch() [ 645.481774] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 645.481774] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] result = function(*args, **kwargs) [ 645.481774] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 645.481774] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] return func(*args, **kwargs) [ 645.481774] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 645.481774] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] raise e [ 645.481774] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 645.481774] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] nwinfo = self.network_api.allocate_for_instance( [ 645.481774] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 645.481774] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] created_port_ids = self._update_ports_for_instance( [ 645.481774] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 645.481774] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] with excutils.save_and_reraise_exception(): [ 645.481774] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.483065] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] self.force_reraise() [ 645.483065] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.483065] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] raise self.value [ 645.483065] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 645.483065] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] updated_port = self._update_port( [ 645.483065] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.483065] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] _ensure_no_port_binding_failure(port) [ 645.483065] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.483065] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] raise exception.PortBindingFailed(port_id=port['id']) [ 645.483065] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] nova.exception.PortBindingFailed: Binding failed for port d33d9daf-8548-4f1d-a09e-ed9348cf5d0c, please check neutron logs for more information. [ 645.483065] env[62692]: ERROR nova.compute.manager [instance: 647d9514-5c31-4970-82e7-626e08f00237] [ 645.483777] env[62692]: DEBUG nova.compute.utils [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Binding failed for port d33d9daf-8548-4f1d-a09e-ed9348cf5d0c, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 645.484767] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 12.406s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.492769] env[62692]: DEBUG nova.compute.manager [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Build of instance 647d9514-5c31-4970-82e7-626e08f00237 was re-scheduled: Binding failed for port d33d9daf-8548-4f1d-a09e-ed9348cf5d0c, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 645.493238] env[62692]: DEBUG nova.compute.manager [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 645.493600] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Acquiring lock "refresh_cache-647d9514-5c31-4970-82e7-626e08f00237" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.493689] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Acquired lock "refresh_cache-647d9514-5c31-4970-82e7-626e08f00237" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.493780] env[62692]: DEBUG nova.network.neutron [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 645.508526] env[62692]: DEBUG nova.network.neutron [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 645.529389] env[62692]: DEBUG nova.network.neutron [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.818010] env[62692]: DEBUG nova.network.neutron [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.863146] env[62692]: DEBUG nova.compute.manager [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 646.031811] env[62692]: INFO nova.compute.manager [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e] Took 1.17 seconds to deallocate network for instance. [ 646.035701] env[62692]: DEBUG nova.network.neutron [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 646.168782] env[62692]: DEBUG nova.network.neutron [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.224034] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Acquiring lock "ed43eb23-0f75-4a8b-96a1-142e1abe400f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.224259] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Lock "ed43eb23-0f75-4a8b-96a1-142e1abe400f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.324187] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Releasing lock "refresh_cache-f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.324187] env[62692]: DEBUG nova.compute.manager [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 646.324187] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 646.324187] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1f5982f9-4b93-4722-b4c8-304202a92040 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.335453] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95340f84-e052-4c22-9ffe-d4f7d67bffd9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.358911] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed could not be found. [ 646.359233] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 646.359505] env[62692]: INFO nova.compute.manager [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Took 0.04 seconds to destroy the instance on the hypervisor. [ 646.359855] env[62692]: DEBUG oslo.service.loopingcall [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 646.360169] env[62692]: DEBUG nova.compute.manager [-] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 646.360355] env[62692]: DEBUG nova.network.neutron [-] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 646.394703] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.396612] env[62692]: DEBUG nova.network.neutron [-] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 646.543094] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Acquiring lock "700d3723-f787-4770-a409-642cd8085e9c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.543376] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Lock "700d3723-f787-4770-a409-642cd8085e9c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.629319] env[62692]: DEBUG nova.compute.manager [req-d273fe0b-63b4-46b7-8cbc-0c84449725b5 req-a22f903e-698c-46e1-9942-2a5bc2006418 service nova] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Received event network-changed-d37419d2-13ae-4fd0-8880-7dd92681d221 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 646.629607] env[62692]: DEBUG nova.compute.manager [req-d273fe0b-63b4-46b7-8cbc-0c84449725b5 req-a22f903e-698c-46e1-9942-2a5bc2006418 service nova] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Refreshing instance network info cache due to event network-changed-d37419d2-13ae-4fd0-8880-7dd92681d221. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 646.630020] env[62692]: DEBUG oslo_concurrency.lockutils [req-d273fe0b-63b4-46b7-8cbc-0c84449725b5 req-a22f903e-698c-46e1-9942-2a5bc2006418 service nova] Acquiring lock "refresh_cache-f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.630020] env[62692]: DEBUG oslo_concurrency.lockutils [req-d273fe0b-63b4-46b7-8cbc-0c84449725b5 req-a22f903e-698c-46e1-9942-2a5bc2006418 service nova] Acquired lock "refresh_cache-f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.630204] env[62692]: DEBUG nova.network.neutron [req-d273fe0b-63b4-46b7-8cbc-0c84449725b5 req-a22f903e-698c-46e1-9942-2a5bc2006418 service nova] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Refreshing network info cache for port d37419d2-13ae-4fd0-8880-7dd92681d221 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 646.671477] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Releasing lock "refresh_cache-647d9514-5c31-4970-82e7-626e08f00237" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.671738] env[62692]: DEBUG nova.compute.manager [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 646.671940] env[62692]: DEBUG nova.compute.manager [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 646.672167] env[62692]: DEBUG nova.network.neutron [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 646.702211] env[62692]: DEBUG nova.network.neutron [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 646.902323] env[62692]: DEBUG nova.network.neutron [-] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.045738] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 647.096134] env[62692]: INFO nova.scheduler.client.report [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Deleted allocations for instance 5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e [ 647.164561] env[62692]: DEBUG nova.network.neutron [req-d273fe0b-63b4-46b7-8cbc-0c84449725b5 req-a22f903e-698c-46e1-9942-2a5bc2006418 service nova] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 647.204883] env[62692]: DEBUG nova.network.neutron [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.298534] env[62692]: DEBUG nova.network.neutron [req-d273fe0b-63b4-46b7-8cbc-0c84449725b5 req-a22f903e-698c-46e1-9942-2a5bc2006418 service nova] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.405656] env[62692]: INFO nova.compute.manager [-] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Took 1.05 seconds to deallocate network for instance. [ 647.409221] env[62692]: DEBUG nova.compute.claims [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 647.409378] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.557320] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 647d9514-5c31-4970-82e7-626e08f00237 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 647.557320] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 4fabba8d-25f5-48a0-a844-bdec6ea5c422 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 647.557320] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 54f615e3-5c6e-44bb-aeb7-5620a639b55c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 647.557320] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 647.608320] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f6bbf462-b547-4391-a566-609c31555f46 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Lock "5dc45cde-78d7-4d00-bcd4-efeb6dd6d29e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.102s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.620884] env[62692]: DEBUG oslo_concurrency.lockutils [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Acquiring lock "49ee893a-313e-42ed-b932-352d450e8645" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.621259] env[62692]: DEBUG oslo_concurrency.lockutils [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Lock "49ee893a-313e-42ed-b932-352d450e8645" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.710296] env[62692]: INFO nova.compute.manager [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] [instance: 647d9514-5c31-4970-82e7-626e08f00237] Took 1.04 seconds to deallocate network for instance. [ 647.801811] env[62692]: DEBUG oslo_concurrency.lockutils [req-d273fe0b-63b4-46b7-8cbc-0c84449725b5 req-a22f903e-698c-46e1-9942-2a5bc2006418 service nova] Releasing lock "refresh_cache-f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.061262] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance e6a36771-c121-4ff5-8a78-c7b7868bf95b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 648.104391] env[62692]: DEBUG nova.compute.manager [None req-d17dab40-4e3a-4f52-9ad2-e65e7a5f71dc tempest-ServerDiagnosticsV248Test-2118826336 tempest-ServerDiagnosticsV248Test-2118826336-project-admin] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 648.106712] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7ccc690-f6ac-4bbd-81f7-5b144c665de5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.111024] env[62692]: DEBUG nova.compute.manager [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 648.120777] env[62692]: INFO nova.compute.manager [None req-d17dab40-4e3a-4f52-9ad2-e65e7a5f71dc tempest-ServerDiagnosticsV248Test-2118826336 tempest-ServerDiagnosticsV248Test-2118826336-project-admin] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Retrieving diagnostics [ 648.122055] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9517b6f1-11d2-4a88-8a39-c730f5296981 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.526259] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "d99737d8-2eb0-40ee-b61e-6c736c84ea59" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.526622] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "d99737d8-2eb0-40ee-b61e-6c736c84ea59" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.565874] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 891170e7-35db-4faa-a1be-fcf376faee93 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 648.658619] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.761877] env[62692]: INFO nova.scheduler.client.report [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Deleted allocations for instance 647d9514-5c31-4970-82e7-626e08f00237 [ 649.070196] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 59882012-b01c-45ec-bd53-6e17e356ce5f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 649.277543] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a0bddc89-61a8-45b8-9aa0-1d2ad7ee40a9 tempest-VolumesAssistedSnapshotsTest-1365700441 tempest-VolumesAssistedSnapshotsTest-1365700441-project-member] Lock "647d9514-5c31-4970-82e7-626e08f00237" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.506s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 649.351424] env[62692]: DEBUG oslo_concurrency.lockutils [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Acquiring lock "628cc505-3edf-4066-91be-da009ebcf219" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.351664] env[62692]: DEBUG oslo_concurrency.lockutils [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Lock "628cc505-3edf-4066-91be-da009ebcf219" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.408250] env[62692]: DEBUG nova.compute.manager [req-c812b60f-07dc-4ec1-a414-ef4c6a4fd114 req-267ccca7-9faf-455e-87ef-c4ba6ca6be07 service nova] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Received event network-vif-deleted-d37419d2-13ae-4fd0-8880-7dd92681d221 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 649.575526] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance b014c45b-4d27-4ade-b823-0dd64dfea2e2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 649.781745] env[62692]: DEBUG nova.compute.manager [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 650.079265] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 6883c19f-1e3a-4eac-be74-8e725faee77a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 650.314717] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 650.587949] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 844b3803-8a01-484e-8908-dbae27188acb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 650.850594] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Acquiring lock "abee88ff-5cf5-4bf1-91e0-93b19cf30046" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 650.850594] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lock "abee88ff-5cf5-4bf1-91e0-93b19cf30046" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.091850] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 0aac5bf9-2af4-4745-a961-46986fb57779 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 651.596264] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 5d18922d-5937-47c8-a04a-aa757c153e8f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 652.100900] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance b76cce5f-b4ab-424f-b604-684f5c22154a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 652.606986] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 9422a9a8-795d-4794-8ba9-9e6f88cdb464 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 653.112644] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 2feffd08-f011-4117-9f8d-ac6730f893f5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 653.618616] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 146e164e-2543-46ab-a7d3-9803d56f61ec has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 654.122062] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance dbd51115-6b41-4019-955d-16a52b534e84 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 654.600056] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "fa7bb883-0065-4ecc-9eb5-8c5443344a2a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.600271] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "fa7bb883-0065-4ecc-9eb5-8c5443344a2a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.625416] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 49d69437-7046-4cb2-914c-8544dca70d3f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 655.131543] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance db727822-af78-49bc-98d9-9caabc35b3b6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 655.636313] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 0266b67a-7c53-471d-a167-ac7362c4b9be has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 656.065188] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Acquiring lock "c1818ed2-9e27-4738-bc17-98832318cf61" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.065188] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Lock "c1818ed2-9e27-4738-bc17-98832318cf61" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.140133] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance a5fa20ab-66b7-4632-8a80-ac031263c042 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 656.645838] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance ed43eb23-0f75-4a8b-96a1-142e1abe400f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 656.851916] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "5a08fa8a-f9fe-4879-bb7b-baa04097df6c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.852256] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "5a08fa8a-f9fe-4879-bb7b-baa04097df6c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.153301] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 700d3723-f787-4770-a409-642cd8085e9c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 657.153667] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62692) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 657.153931] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1152MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62692) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 657.681311] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac416130-1f1d-4bf9-b400-96a3ca8645af {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.689954] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7314043a-150a-470d-910b-dfc9d41f709d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.728313] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a62f80d-0905-4806-94bc-b259e58689ec {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.737853] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-941cfaae-df6c-49af-b59d-fde129cdb730 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.752969] env[62692]: DEBUG nova.compute.provider_tree [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.256657] env[62692]: DEBUG nova.scheduler.client.report [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 658.683782] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "9b8e830c-61b7-4dd2-8324-d3a96eec1465" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.683782] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "9b8e830c-61b7-4dd2-8324-d3a96eec1465" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 658.763930] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62692) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 658.763930] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 13.278s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.763930] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.155s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 658.764799] env[62692]: INFO nova.compute.claims [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 659.035560] env[62692]: DEBUG nova.compute.manager [None req-674b2f0a-c3ed-4bab-9261-72041a78de7e tempest-ServerDiagnosticsV248Test-2118826336 tempest-ServerDiagnosticsV248Test-2118826336-project-admin] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 659.037808] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e1ac181-dcb4-4f7b-ae06-3108f1764fe9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.048116] env[62692]: INFO nova.compute.manager [None req-674b2f0a-c3ed-4bab-9261-72041a78de7e tempest-ServerDiagnosticsV248Test-2118826336 tempest-ServerDiagnosticsV248Test-2118826336-project-admin] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Retrieving diagnostics [ 659.050094] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a07475-490a-4779-b55d-5a1f6516c926 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.577885] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Acquiring lock "105e6fd5-4eff-4a0c-9a4b-e6deade781d5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.578258] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Lock "105e6fd5-4eff-4a0c-9a4b-e6deade781d5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.371373] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50a16591-57a8-409f-a79c-9fd09f35e00d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.382463] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f10f1a5-86cb-4387-a113-602a723d486b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.412379] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd1e4aab-242f-4973-b19b-4a481c6e24eb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.420174] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f8b30c8-56ab-4152-bba2-abf47aff5536 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.434389] env[62692]: DEBUG nova.compute.provider_tree [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 660.660693] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "bb547773-d176-4c8e-a0fa-a374d5050b1b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.660995] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "bb547773-d176-4c8e-a0fa-a374d5050b1b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.783107] env[62692]: DEBUG oslo_concurrency.lockutils [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Acquiring lock "54f615e3-5c6e-44bb-aeb7-5620a639b55c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.783107] env[62692]: DEBUG oslo_concurrency.lockutils [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Lock "54f615e3-5c6e-44bb-aeb7-5620a639b55c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.783107] env[62692]: DEBUG oslo_concurrency.lockutils [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Acquiring lock "54f615e3-5c6e-44bb-aeb7-5620a639b55c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.783107] env[62692]: DEBUG oslo_concurrency.lockutils [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Lock "54f615e3-5c6e-44bb-aeb7-5620a639b55c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.783638] env[62692]: DEBUG oslo_concurrency.lockutils [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Lock "54f615e3-5c6e-44bb-aeb7-5620a639b55c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.788704] env[62692]: INFO nova.compute.manager [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Terminating instance [ 660.793630] env[62692]: DEBUG oslo_concurrency.lockutils [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Acquiring lock "refresh_cache-54f615e3-5c6e-44bb-aeb7-5620a639b55c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.793913] env[62692]: DEBUG oslo_concurrency.lockutils [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Acquired lock "refresh_cache-54f615e3-5c6e-44bb-aeb7-5620a639b55c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.794362] env[62692]: DEBUG nova.network.neutron [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 660.938526] env[62692]: DEBUG nova.scheduler.client.report [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 661.317792] env[62692]: DEBUG nova.network.neutron [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 661.436668] env[62692]: DEBUG nova.network.neutron [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.443731] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.681s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.444430] env[62692]: DEBUG nova.compute.manager [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 661.447291] env[62692]: DEBUG oslo_concurrency.lockutils [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.810s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.448810] env[62692]: INFO nova.compute.claims [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 661.944728] env[62692]: DEBUG oslo_concurrency.lockutils [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Releasing lock "refresh_cache-54f615e3-5c6e-44bb-aeb7-5620a639b55c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.945201] env[62692]: DEBUG nova.compute.manager [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 661.945484] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 661.946440] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-827a2c29-32d7-48da-8fe5-614cba195b3b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.953622] env[62692]: DEBUG nova.compute.utils [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 661.964534] env[62692]: DEBUG nova.compute.manager [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 661.964718] env[62692]: DEBUG nova.network.neutron [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 661.967314] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 661.976390] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f0dfd899-1ddc-474a-9f0c-6026e94e5a44 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.978440] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.979039] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.985146] env[62692]: DEBUG oslo_vmware.api [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Waiting for the task: (returnval){ [ 661.985146] env[62692]: value = "task-1140888" [ 661.985146] env[62692]: _type = "Task" [ 661.985146] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.994763] env[62692]: DEBUG oslo_vmware.api [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': task-1140888, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.016539] env[62692]: DEBUG nova.policy [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '409c9e69722b434bb2c97bfcd17c8877', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bfc6add89b3e416e8d1d6a45d930e461', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 662.422198] env[62692]: DEBUG oslo_concurrency.lockutils [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Acquiring lock "6097d6b2-52d5-4765-94c0-d5f3609464d8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.422198] env[62692]: DEBUG oslo_concurrency.lockutils [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Lock "6097d6b2-52d5-4765-94c0-d5f3609464d8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.462478] env[62692]: DEBUG nova.compute.manager [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 662.500323] env[62692]: DEBUG oslo_vmware.api [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': task-1140888, 'name': PowerOffVM_Task, 'duration_secs': 0.111283} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.501472] env[62692]: DEBUG nova.network.neutron [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Successfully created port: 4122978e-21d6-422d-a286-589614e44e7c {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 662.504376] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 662.504562] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 662.504828] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-465fad60-07e2-41fe-9328-4a34abb5b922 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.536474] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 662.536686] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 662.536870] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Deleting the datastore file [datastore2] 54f615e3-5c6e-44bb-aeb7-5620a639b55c {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 662.537138] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d06b1bac-991c-4322-8d0d-c9289e08e105 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.545202] env[62692]: DEBUG oslo_vmware.api [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Waiting for the task: (returnval){ [ 662.545202] env[62692]: value = "task-1140890" [ 662.545202] env[62692]: _type = "Task" [ 662.545202] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.560982] env[62692]: DEBUG oslo_vmware.api [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': task-1140890, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.063745] env[62692]: DEBUG oslo_vmware.api [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Task: {'id': task-1140890, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.10147} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.064169] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 663.064260] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 663.064372] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 663.064582] env[62692]: INFO nova.compute.manager [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Took 1.12 seconds to destroy the instance on the hypervisor. [ 663.064830] env[62692]: DEBUG oslo.service.loopingcall [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 663.065126] env[62692]: DEBUG nova.compute.manager [-] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 663.065126] env[62692]: DEBUG nova.network.neutron [-] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 663.085316] env[62692]: DEBUG nova.network.neutron [-] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 663.093254] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca68dd6c-465e-402a-9259-8af0e3fbd9d0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.102539] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67b2f62d-9e48-4b44-bef1-ddf48948e7d4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.143900] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65ed2091-4b46-41dd-9404-32620ad4ac8e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.158969] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2858a55-8916-41ae-96dc-d0fc41bc64f4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.174853] env[62692]: DEBUG nova.compute.provider_tree [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 663.477896] env[62692]: DEBUG nova.compute.manager [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 663.509134] env[62692]: DEBUG nova.virt.hardware [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 663.509300] env[62692]: DEBUG nova.virt.hardware [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 663.509525] env[62692]: DEBUG nova.virt.hardware [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 663.509619] env[62692]: DEBUG nova.virt.hardware [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 663.509775] env[62692]: DEBUG nova.virt.hardware [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 663.509917] env[62692]: DEBUG nova.virt.hardware [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 663.510707] env[62692]: DEBUG nova.virt.hardware [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 663.510707] env[62692]: DEBUG nova.virt.hardware [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 663.510843] env[62692]: DEBUG nova.virt.hardware [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 663.511468] env[62692]: DEBUG nova.virt.hardware [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 663.511468] env[62692]: DEBUG nova.virt.hardware [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 663.512406] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6455473-776c-46c0-a9c0-167a3d24f047 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.521599] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8812bdd-e592-4b49-a7eb-54c19453c29d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.588652] env[62692]: DEBUG nova.network.neutron [-] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.681852] env[62692]: DEBUG nova.scheduler.client.report [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 664.094332] env[62692]: INFO nova.compute.manager [-] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Took 1.03 seconds to deallocate network for instance. [ 664.188947] env[62692]: DEBUG oslo_concurrency.lockutils [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.739s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.188947] env[62692]: DEBUG nova.compute.manager [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 664.192932] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.604s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.601484] env[62692]: DEBUG oslo_concurrency.lockutils [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.702204] env[62692]: DEBUG nova.compute.utils [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 664.706976] env[62692]: DEBUG nova.compute.manager [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 664.708727] env[62692]: DEBUG nova.network.neutron [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 664.817047] env[62692]: DEBUG nova.policy [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd1699d57726d4bff8a6b59acdcc00a60', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '109b083810e248b5adca51a6cc5e377e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 664.990262] env[62692]: DEBUG nova.compute.manager [req-0002bc0a-5d8e-45dd-973c-fa214cc85a05 req-c244b3aa-7444-4b82-9236-08c0eca474a2 service nova] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Received event network-changed-4122978e-21d6-422d-a286-589614e44e7c {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 664.990262] env[62692]: DEBUG nova.compute.manager [req-0002bc0a-5d8e-45dd-973c-fa214cc85a05 req-c244b3aa-7444-4b82-9236-08c0eca474a2 service nova] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Refreshing instance network info cache due to event network-changed-4122978e-21d6-422d-a286-589614e44e7c. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 664.990434] env[62692]: DEBUG oslo_concurrency.lockutils [req-0002bc0a-5d8e-45dd-973c-fa214cc85a05 req-c244b3aa-7444-4b82-9236-08c0eca474a2 service nova] Acquiring lock "refresh_cache-e6a36771-c121-4ff5-8a78-c7b7868bf95b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.990505] env[62692]: DEBUG oslo_concurrency.lockutils [req-0002bc0a-5d8e-45dd-973c-fa214cc85a05 req-c244b3aa-7444-4b82-9236-08c0eca474a2 service nova] Acquired lock "refresh_cache-e6a36771-c121-4ff5-8a78-c7b7868bf95b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.991337] env[62692]: DEBUG nova.network.neutron [req-0002bc0a-5d8e-45dd-973c-fa214cc85a05 req-c244b3aa-7444-4b82-9236-08c0eca474a2 service nova] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Refreshing network info cache for port 4122978e-21d6-422d-a286-589614e44e7c {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 665.208563] env[62692]: DEBUG nova.compute.manager [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 665.269872] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-952dd671-51e5-45e8-9852-12e5165ee8e8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.272841] env[62692]: DEBUG nova.network.neutron [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Successfully created port: e58be796-e8e0-4b09-a0bf-334afc678ee1 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 665.279484] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aee8f28c-35ee-4730-9e21-31e8865b7327 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.313202] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3fb50de-a2a8-46db-be5f-50795bae44b6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.321025] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-816f7846-3fe1-4990-bf27-2aaa82dbd09f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.335615] env[62692]: DEBUG nova.compute.provider_tree [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 665.337859] env[62692]: ERROR nova.compute.manager [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4122978e-21d6-422d-a286-589614e44e7c, please check neutron logs for more information. [ 665.337859] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 665.337859] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.337859] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 665.337859] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 665.337859] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 665.337859] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 665.337859] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 665.337859] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.337859] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 665.337859] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.337859] env[62692]: ERROR nova.compute.manager raise self.value [ 665.337859] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 665.337859] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 665.337859] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.337859] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 665.338407] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.338407] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 665.338407] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4122978e-21d6-422d-a286-589614e44e7c, please check neutron logs for more information. [ 665.338407] env[62692]: ERROR nova.compute.manager [ 665.338407] env[62692]: Traceback (most recent call last): [ 665.338407] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 665.338407] env[62692]: listener.cb(fileno) [ 665.338407] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.338407] env[62692]: result = function(*args, **kwargs) [ 665.338407] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 665.338407] env[62692]: return func(*args, **kwargs) [ 665.338407] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.338407] env[62692]: raise e [ 665.338407] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.338407] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 665.338407] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 665.338407] env[62692]: created_port_ids = self._update_ports_for_instance( [ 665.338407] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 665.338407] env[62692]: with excutils.save_and_reraise_exception(): [ 665.338407] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.338407] env[62692]: self.force_reraise() [ 665.338407] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.338407] env[62692]: raise self.value [ 665.338407] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 665.338407] env[62692]: updated_port = self._update_port( [ 665.338407] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.338407] env[62692]: _ensure_no_port_binding_failure(port) [ 665.338407] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.338407] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 665.339402] env[62692]: nova.exception.PortBindingFailed: Binding failed for port 4122978e-21d6-422d-a286-589614e44e7c, please check neutron logs for more information. [ 665.339402] env[62692]: Removing descriptor: 15 [ 665.339402] env[62692]: ERROR nova.compute.manager [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4122978e-21d6-422d-a286-589614e44e7c, please check neutron logs for more information. [ 665.339402] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Traceback (most recent call last): [ 665.339402] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 665.339402] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] yield resources [ 665.339402] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 665.339402] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] self.driver.spawn(context, instance, image_meta, [ 665.339402] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 665.339402] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 665.339402] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 665.339402] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] vm_ref = self.build_virtual_machine(instance, [ 665.339762] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 665.339762] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] vif_infos = vmwarevif.get_vif_info(self._session, [ 665.339762] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 665.339762] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] for vif in network_info: [ 665.339762] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 665.339762] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] return self._sync_wrapper(fn, *args, **kwargs) [ 665.339762] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 665.339762] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] self.wait() [ 665.339762] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 665.339762] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] self[:] = self._gt.wait() [ 665.339762] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 665.339762] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] return self._exit_event.wait() [ 665.339762] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 665.340152] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] result = hub.switch() [ 665.340152] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 665.340152] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] return self.greenlet.switch() [ 665.340152] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.340152] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] result = function(*args, **kwargs) [ 665.340152] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 665.340152] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] return func(*args, **kwargs) [ 665.340152] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.340152] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] raise e [ 665.340152] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.340152] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] nwinfo = self.network_api.allocate_for_instance( [ 665.340152] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 665.340152] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] created_port_ids = self._update_ports_for_instance( [ 665.340613] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 665.340613] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] with excutils.save_and_reraise_exception(): [ 665.340613] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.340613] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] self.force_reraise() [ 665.340613] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.340613] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] raise self.value [ 665.340613] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 665.340613] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] updated_port = self._update_port( [ 665.340613] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.340613] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] _ensure_no_port_binding_failure(port) [ 665.340613] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.340613] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] raise exception.PortBindingFailed(port_id=port['id']) [ 665.341743] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] nova.exception.PortBindingFailed: Binding failed for port 4122978e-21d6-422d-a286-589614e44e7c, please check neutron logs for more information. [ 665.341743] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] [ 665.341743] env[62692]: INFO nova.compute.manager [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Terminating instance [ 665.341743] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Acquiring lock "refresh_cache-e6a36771-c121-4ff5-8a78-c7b7868bf95b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.539020] env[62692]: DEBUG nova.network.neutron [req-0002bc0a-5d8e-45dd-973c-fa214cc85a05 req-c244b3aa-7444-4b82-9236-08c0eca474a2 service nova] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 665.841889] env[62692]: DEBUG nova.scheduler.client.report [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 665.846282] env[62692]: DEBUG nova.network.neutron [req-0002bc0a-5d8e-45dd-973c-fa214cc85a05 req-c244b3aa-7444-4b82-9236-08c0eca474a2 service nova] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.224297] env[62692]: DEBUG nova.compute.manager [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 666.260633] env[62692]: DEBUG nova.virt.hardware [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:29:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1537058302',id=21,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1157775648',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 666.260889] env[62692]: DEBUG nova.virt.hardware [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 666.261070] env[62692]: DEBUG nova.virt.hardware [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 666.261308] env[62692]: DEBUG nova.virt.hardware [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 666.262078] env[62692]: DEBUG nova.virt.hardware [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 666.262277] env[62692]: DEBUG nova.virt.hardware [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 666.262501] env[62692]: DEBUG nova.virt.hardware [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 666.262665] env[62692]: DEBUG nova.virt.hardware [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 666.262832] env[62692]: DEBUG nova.virt.hardware [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 666.262994] env[62692]: DEBUG nova.virt.hardware [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 666.263185] env[62692]: DEBUG nova.virt.hardware [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 666.264056] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84bfc7ea-012c-41d0-8024-340c24b3bcb5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.275954] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0982a64d-591b-44e5-a702-e4e52272a921 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.346557] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.155s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.347231] env[62692]: ERROR nova.compute.manager [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port dc5d38ec-06fc-440f-bd69-0a197ea6da73, please check neutron logs for more information. [ 666.347231] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Traceback (most recent call last): [ 666.347231] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 666.347231] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] self.driver.spawn(context, instance, image_meta, [ 666.347231] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 666.347231] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] self._vmops.spawn(context, instance, image_meta, injected_files, [ 666.347231] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 666.347231] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] vm_ref = self.build_virtual_machine(instance, [ 666.347231] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 666.347231] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] vif_infos = vmwarevif.get_vif_info(self._session, [ 666.347231] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 666.347638] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] for vif in network_info: [ 666.347638] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 666.347638] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] return self._sync_wrapper(fn, *args, **kwargs) [ 666.347638] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 666.347638] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] self.wait() [ 666.347638] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 666.347638] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] self[:] = self._gt.wait() [ 666.347638] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 666.347638] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] return self._exit_event.wait() [ 666.347638] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 666.347638] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] result = hub.switch() [ 666.347638] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 666.347638] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] return self.greenlet.switch() [ 666.348049] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 666.348049] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] result = function(*args, **kwargs) [ 666.348049] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 666.348049] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] return func(*args, **kwargs) [ 666.348049] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 666.348049] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] raise e [ 666.348049] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 666.348049] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] nwinfo = self.network_api.allocate_for_instance( [ 666.348049] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 666.348049] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] created_port_ids = self._update_ports_for_instance( [ 666.348049] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 666.348049] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] with excutils.save_and_reraise_exception(): [ 666.348049] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.348513] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] self.force_reraise() [ 666.348513] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.348513] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] raise self.value [ 666.348513] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 666.348513] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] updated_port = self._update_port( [ 666.348513] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.348513] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] _ensure_no_port_binding_failure(port) [ 666.348513] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.348513] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] raise exception.PortBindingFailed(port_id=port['id']) [ 666.348513] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] nova.exception.PortBindingFailed: Binding failed for port dc5d38ec-06fc-440f-bd69-0a197ea6da73, please check neutron logs for more information. [ 666.348513] env[62692]: ERROR nova.compute.manager [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] [ 666.348857] env[62692]: DEBUG nova.compute.utils [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Binding failed for port dc5d38ec-06fc-440f-bd69-0a197ea6da73, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 666.349385] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.279s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.350945] env[62692]: INFO nova.compute.claims [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 666.353502] env[62692]: DEBUG oslo_concurrency.lockutils [req-0002bc0a-5d8e-45dd-973c-fa214cc85a05 req-c244b3aa-7444-4b82-9236-08c0eca474a2 service nova] Releasing lock "refresh_cache-e6a36771-c121-4ff5-8a78-c7b7868bf95b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.354120] env[62692]: DEBUG nova.compute.manager [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Build of instance 4fabba8d-25f5-48a0-a844-bdec6ea5c422 was re-scheduled: Binding failed for port dc5d38ec-06fc-440f-bd69-0a197ea6da73, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 666.354570] env[62692]: DEBUG nova.compute.manager [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 666.354793] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Acquiring lock "refresh_cache-4fabba8d-25f5-48a0-a844-bdec6ea5c422" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 666.354936] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Acquired lock "refresh_cache-4fabba8d-25f5-48a0-a844-bdec6ea5c422" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.355108] env[62692]: DEBUG nova.network.neutron [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 666.357183] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Acquired lock "refresh_cache-e6a36771-c121-4ff5-8a78-c7b7868bf95b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.357385] env[62692]: DEBUG nova.network.neutron [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 666.888881] env[62692]: DEBUG nova.network.neutron [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 666.899881] env[62692]: DEBUG nova.network.neutron [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 666.990362] env[62692]: DEBUG nova.network.neutron [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.057571] env[62692]: DEBUG nova.compute.manager [req-4ecc1ced-3fa7-42c4-8d5b-a3af5e9cc6bf req-903802ca-0eab-478a-b471-960ccd91d9ef service nova] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Received event network-vif-deleted-4122978e-21d6-422d-a286-589614e44e7c {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 667.103542] env[62692]: DEBUG nova.network.neutron [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.210335] env[62692]: DEBUG nova.compute.manager [req-9301f8d4-2ca3-4045-814a-01e178f09737 req-c504209d-94ff-4548-a484-65918f1cb726 service nova] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Received event network-changed-e58be796-e8e0-4b09-a0bf-334afc678ee1 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 667.210335] env[62692]: DEBUG nova.compute.manager [req-9301f8d4-2ca3-4045-814a-01e178f09737 req-c504209d-94ff-4548-a484-65918f1cb726 service nova] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Refreshing instance network info cache due to event network-changed-e58be796-e8e0-4b09-a0bf-334afc678ee1. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 667.210335] env[62692]: DEBUG oslo_concurrency.lockutils [req-9301f8d4-2ca3-4045-814a-01e178f09737 req-c504209d-94ff-4548-a484-65918f1cb726 service nova] Acquiring lock "refresh_cache-891170e7-35db-4faa-a1be-fcf376faee93" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.210572] env[62692]: DEBUG oslo_concurrency.lockutils [req-9301f8d4-2ca3-4045-814a-01e178f09737 req-c504209d-94ff-4548-a484-65918f1cb726 service nova] Acquired lock "refresh_cache-891170e7-35db-4faa-a1be-fcf376faee93" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.211150] env[62692]: DEBUG nova.network.neutron [req-9301f8d4-2ca3-4045-814a-01e178f09737 req-c504209d-94ff-4548-a484-65918f1cb726 service nova] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Refreshing network info cache for port e58be796-e8e0-4b09-a0bf-334afc678ee1 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 667.385448] env[62692]: ERROR nova.compute.manager [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e58be796-e8e0-4b09-a0bf-334afc678ee1, please check neutron logs for more information. [ 667.385448] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 667.385448] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 667.385448] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 667.385448] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 667.385448] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 667.385448] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 667.385448] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 667.385448] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.385448] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 667.385448] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.385448] env[62692]: ERROR nova.compute.manager raise self.value [ 667.385448] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 667.385448] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 667.385448] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.385448] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 667.385957] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.385957] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 667.385957] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e58be796-e8e0-4b09-a0bf-334afc678ee1, please check neutron logs for more information. [ 667.385957] env[62692]: ERROR nova.compute.manager [ 667.385957] env[62692]: Traceback (most recent call last): [ 667.385957] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 667.385957] env[62692]: listener.cb(fileno) [ 667.385957] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 667.385957] env[62692]: result = function(*args, **kwargs) [ 667.385957] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 667.385957] env[62692]: return func(*args, **kwargs) [ 667.385957] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 667.385957] env[62692]: raise e [ 667.385957] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 667.385957] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 667.385957] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 667.385957] env[62692]: created_port_ids = self._update_ports_for_instance( [ 667.385957] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 667.385957] env[62692]: with excutils.save_and_reraise_exception(): [ 667.385957] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.385957] env[62692]: self.force_reraise() [ 667.385957] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.385957] env[62692]: raise self.value [ 667.385957] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 667.385957] env[62692]: updated_port = self._update_port( [ 667.385957] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.385957] env[62692]: _ensure_no_port_binding_failure(port) [ 667.385957] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.385957] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 667.386815] env[62692]: nova.exception.PortBindingFailed: Binding failed for port e58be796-e8e0-4b09-a0bf-334afc678ee1, please check neutron logs for more information. [ 667.386815] env[62692]: Removing descriptor: 19 [ 667.386815] env[62692]: ERROR nova.compute.manager [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e58be796-e8e0-4b09-a0bf-334afc678ee1, please check neutron logs for more information. [ 667.386815] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Traceback (most recent call last): [ 667.386815] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 667.386815] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] yield resources [ 667.386815] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 667.386815] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] self.driver.spawn(context, instance, image_meta, [ 667.386815] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 667.386815] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] self._vmops.spawn(context, instance, image_meta, injected_files, [ 667.386815] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 667.386815] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] vm_ref = self.build_virtual_machine(instance, [ 667.387182] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 667.387182] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] vif_infos = vmwarevif.get_vif_info(self._session, [ 667.387182] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 667.387182] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] for vif in network_info: [ 667.387182] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 667.387182] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] return self._sync_wrapper(fn, *args, **kwargs) [ 667.387182] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 667.387182] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] self.wait() [ 667.387182] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 667.387182] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] self[:] = self._gt.wait() [ 667.387182] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 667.387182] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] return self._exit_event.wait() [ 667.387182] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 667.394502] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] result = hub.switch() [ 667.394502] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 667.394502] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] return self.greenlet.switch() [ 667.394502] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 667.394502] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] result = function(*args, **kwargs) [ 667.394502] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 667.394502] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] return func(*args, **kwargs) [ 667.394502] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 667.394502] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] raise e [ 667.394502] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 667.394502] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] nwinfo = self.network_api.allocate_for_instance( [ 667.394502] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 667.394502] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] created_port_ids = self._update_ports_for_instance( [ 667.395136] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 667.395136] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] with excutils.save_and_reraise_exception(): [ 667.395136] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.395136] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] self.force_reraise() [ 667.395136] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.395136] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] raise self.value [ 667.395136] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 667.395136] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] updated_port = self._update_port( [ 667.395136] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.395136] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] _ensure_no_port_binding_failure(port) [ 667.395136] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.395136] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] raise exception.PortBindingFailed(port_id=port['id']) [ 667.395543] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] nova.exception.PortBindingFailed: Binding failed for port e58be796-e8e0-4b09-a0bf-334afc678ee1, please check neutron logs for more information. [ 667.395543] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] [ 667.395543] env[62692]: INFO nova.compute.manager [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Terminating instance [ 667.395543] env[62692]: DEBUG oslo_concurrency.lockutils [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Acquiring lock "refresh_cache-891170e7-35db-4faa-a1be-fcf376faee93" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.494513] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Releasing lock "refresh_cache-4fabba8d-25f5-48a0-a844-bdec6ea5c422" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.494740] env[62692]: DEBUG nova.compute.manager [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 667.497911] env[62692]: DEBUG nova.compute.manager [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 667.497911] env[62692]: DEBUG nova.network.neutron [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 667.523876] env[62692]: DEBUG nova.network.neutron [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 667.608142] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Releasing lock "refresh_cache-e6a36771-c121-4ff5-8a78-c7b7868bf95b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.608142] env[62692]: DEBUG nova.compute.manager [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 667.608142] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 667.608344] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-732b0774-55b2-4514-a492-92465309d037 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.627025] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baab033f-fe53-4088-8d9e-a6944e39ae18 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.650230] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e6a36771-c121-4ff5-8a78-c7b7868bf95b could not be found. [ 667.650873] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 667.650873] env[62692]: INFO nova.compute.manager [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 667.650999] env[62692]: DEBUG oslo.service.loopingcall [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 667.651200] env[62692]: DEBUG nova.compute.manager [-] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 667.651292] env[62692]: DEBUG nova.network.neutron [-] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 667.676088] env[62692]: DEBUG nova.network.neutron [-] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 667.744321] env[62692]: DEBUG nova.network.neutron [req-9301f8d4-2ca3-4045-814a-01e178f09737 req-c504209d-94ff-4548-a484-65918f1cb726 service nova] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 667.899638] env[62692]: DEBUG nova.network.neutron [req-9301f8d4-2ca3-4045-814a-01e178f09737 req-c504209d-94ff-4548-a484-65918f1cb726 service nova] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.012628] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20e6b763-cae8-42b0-8937-e9a2908d3faf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.021198] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d1ba84-e6ea-4b04-a637-a78aa01640de {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.058025] env[62692]: DEBUG nova.network.neutron [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.058240] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dc6112d-669c-4f6b-80ba-e9ff1370ff38 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.065986] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22a5b8f8-3bec-4bd4-b90e-41ab1c281c7b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.080593] env[62692]: DEBUG nova.compute.provider_tree [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 668.184589] env[62692]: DEBUG nova.network.neutron [-] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.298893] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "27583cb4-0478-44d8-8970-478682271513" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.299162] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "27583cb4-0478-44d8-8970-478682271513" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.403668] env[62692]: DEBUG oslo_concurrency.lockutils [req-9301f8d4-2ca3-4045-814a-01e178f09737 req-c504209d-94ff-4548-a484-65918f1cb726 service nova] Releasing lock "refresh_cache-891170e7-35db-4faa-a1be-fcf376faee93" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.404453] env[62692]: DEBUG oslo_concurrency.lockutils [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Acquired lock "refresh_cache-891170e7-35db-4faa-a1be-fcf376faee93" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.404858] env[62692]: DEBUG nova.network.neutron [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 668.562847] env[62692]: INFO nova.compute.manager [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: 4fabba8d-25f5-48a0-a844-bdec6ea5c422] Took 1.07 seconds to deallocate network for instance. [ 668.585701] env[62692]: DEBUG nova.scheduler.client.report [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 668.691145] env[62692]: INFO nova.compute.manager [-] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Took 1.04 seconds to deallocate network for instance. [ 668.693412] env[62692]: DEBUG nova.compute.claims [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 668.694090] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.930784] env[62692]: DEBUG nova.network.neutron [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 669.048064] env[62692]: DEBUG nova.network.neutron [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.093786] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.742s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.093786] env[62692]: DEBUG nova.compute.manager [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 669.095141] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.170s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.100100] env[62692]: INFO nova.compute.claims [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 669.261024] env[62692]: DEBUG nova.compute.manager [req-f26afebf-78b2-45ab-8269-7b07a61650c5 req-f4be5f54-dde3-430f-a68b-3dc2f9f2ef67 service nova] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Received event network-vif-deleted-e58be796-e8e0-4b09-a0bf-334afc678ee1 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 669.549427] env[62692]: DEBUG oslo_concurrency.lockutils [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Releasing lock "refresh_cache-891170e7-35db-4faa-a1be-fcf376faee93" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 669.549917] env[62692]: DEBUG nova.compute.manager [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 669.550962] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 669.551429] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b26d630a-b64e-418c-a81d-93a03d568718 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.563732] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f8badab-607d-45a1-9e1e-a09caafd5b52 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.597295] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 891170e7-35db-4faa-a1be-fcf376faee93 could not be found. [ 669.597510] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 669.597691] env[62692]: INFO nova.compute.manager [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Took 0.05 seconds to destroy the instance on the hypervisor. [ 669.597943] env[62692]: DEBUG oslo.service.loopingcall [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 669.599784] env[62692]: DEBUG nova.compute.manager [-] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 669.599784] env[62692]: DEBUG nova.network.neutron [-] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 669.607338] env[62692]: DEBUG nova.compute.utils [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 669.610966] env[62692]: INFO nova.scheduler.client.report [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Deleted allocations for instance 4fabba8d-25f5-48a0-a844-bdec6ea5c422 [ 669.616445] env[62692]: DEBUG nova.compute.manager [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 669.616445] env[62692]: DEBUG nova.network.neutron [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 669.623754] env[62692]: DEBUG nova.network.neutron [-] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 669.694517] env[62692]: DEBUG nova.policy [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5ba0c537d81444f18ee93c69c9a5cb4f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '95c43089a4c24269aa8f27eb080e69cd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 670.121348] env[62692]: DEBUG nova.compute.manager [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 670.126296] env[62692]: DEBUG nova.network.neutron [-] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.128128] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96161afd-a147-4c94-870d-713c408f6495 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Lock "4fabba8d-25f5-48a0-a844-bdec6ea5c422" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.094s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.181152] env[62692]: DEBUG nova.network.neutron [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Successfully created port: 44e32753-d6cd-4e32-8b26-9476d40e482e {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 670.627822] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aa9640b-200b-4b2c-ae3f-573abc55e9e9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.636124] env[62692]: INFO nova.compute.manager [-] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Took 1.04 seconds to deallocate network for instance. [ 670.636465] env[62692]: DEBUG nova.compute.manager [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 670.647212] env[62692]: DEBUG nova.compute.claims [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 670.647212] env[62692]: DEBUG oslo_concurrency.lockutils [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.648698] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1f93ab5-d003-4682-9a44-11f2cd8ef2f9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.681049] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-080f0750-d632-4b9c-8781-406be68b9bda {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.690760] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc10a92a-6ed0-4160-bac8-d5f6d8f60802 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.707016] env[62692]: DEBUG nova.compute.provider_tree [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 671.145020] env[62692]: DEBUG nova.compute.manager [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 671.164854] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.179548] env[62692]: DEBUG nova.virt.hardware [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 671.179710] env[62692]: DEBUG nova.virt.hardware [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 671.180947] env[62692]: DEBUG nova.virt.hardware [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 671.180947] env[62692]: DEBUG nova.virt.hardware [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 671.180947] env[62692]: DEBUG nova.virt.hardware [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 671.180947] env[62692]: DEBUG nova.virt.hardware [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 671.180947] env[62692]: DEBUG nova.virt.hardware [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 671.181209] env[62692]: DEBUG nova.virt.hardware [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 671.181209] env[62692]: DEBUG nova.virt.hardware [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 671.181209] env[62692]: DEBUG nova.virt.hardware [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 671.181209] env[62692]: DEBUG nova.virt.hardware [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 671.183775] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aba91fb7-fdb4-4564-89f5-8f700ec50e33 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.190654] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25b58630-cff5-43ca-ac75-9649a58ba207 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.209055] env[62692]: DEBUG nova.scheduler.client.report [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 671.637479] env[62692]: DEBUG nova.compute.manager [req-20824974-06ce-4110-a155-ea17f8b56ba1 req-f43bb296-665d-43f8-aaba-632c1d7cb177 service nova] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Received event network-changed-44e32753-d6cd-4e32-8b26-9476d40e482e {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 671.637778] env[62692]: DEBUG nova.compute.manager [req-20824974-06ce-4110-a155-ea17f8b56ba1 req-f43bb296-665d-43f8-aaba-632c1d7cb177 service nova] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Refreshing instance network info cache due to event network-changed-44e32753-d6cd-4e32-8b26-9476d40e482e. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 671.638203] env[62692]: DEBUG oslo_concurrency.lockutils [req-20824974-06ce-4110-a155-ea17f8b56ba1 req-f43bb296-665d-43f8-aaba-632c1d7cb177 service nova] Acquiring lock "refresh_cache-59882012-b01c-45ec-bd53-6e17e356ce5f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.638343] env[62692]: DEBUG oslo_concurrency.lockutils [req-20824974-06ce-4110-a155-ea17f8b56ba1 req-f43bb296-665d-43f8-aaba-632c1d7cb177 service nova] Acquired lock "refresh_cache-59882012-b01c-45ec-bd53-6e17e356ce5f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.638802] env[62692]: DEBUG nova.network.neutron [req-20824974-06ce-4110-a155-ea17f8b56ba1 req-f43bb296-665d-43f8-aaba-632c1d7cb177 service nova] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Refreshing network info cache for port 44e32753-d6cd-4e32-8b26-9476d40e482e {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 671.715643] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.620s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.716630] env[62692]: DEBUG nova.compute.manager [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 671.720681] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.748s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.722238] env[62692]: INFO nova.compute.claims [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 671.816497] env[62692]: ERROR nova.compute.manager [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 44e32753-d6cd-4e32-8b26-9476d40e482e, please check neutron logs for more information. [ 671.816497] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 671.816497] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 671.816497] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 671.816497] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 671.816497] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 671.816497] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 671.816497] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 671.816497] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.816497] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 671.816497] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.816497] env[62692]: ERROR nova.compute.manager raise self.value [ 671.816497] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 671.816497] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 671.816497] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.816497] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 671.817071] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.817071] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 671.817071] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 44e32753-d6cd-4e32-8b26-9476d40e482e, please check neutron logs for more information. [ 671.817071] env[62692]: ERROR nova.compute.manager [ 671.817708] env[62692]: Traceback (most recent call last): [ 671.817708] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 671.817708] env[62692]: listener.cb(fileno) [ 671.817708] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 671.817708] env[62692]: result = function(*args, **kwargs) [ 671.817708] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 671.817708] env[62692]: return func(*args, **kwargs) [ 671.817708] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 671.817708] env[62692]: raise e [ 671.817708] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 671.817708] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 671.817708] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 671.817708] env[62692]: created_port_ids = self._update_ports_for_instance( [ 671.817708] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 671.817708] env[62692]: with excutils.save_and_reraise_exception(): [ 671.817708] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.817708] env[62692]: self.force_reraise() [ 671.817708] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.817708] env[62692]: raise self.value [ 671.817708] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 671.817708] env[62692]: updated_port = self._update_port( [ 671.817708] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.817708] env[62692]: _ensure_no_port_binding_failure(port) [ 671.817708] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.817708] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 671.817708] env[62692]: nova.exception.PortBindingFailed: Binding failed for port 44e32753-d6cd-4e32-8b26-9476d40e482e, please check neutron logs for more information. [ 671.817708] env[62692]: Removing descriptor: 17 [ 671.818929] env[62692]: ERROR nova.compute.manager [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 44e32753-d6cd-4e32-8b26-9476d40e482e, please check neutron logs for more information. [ 671.818929] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Traceback (most recent call last): [ 671.818929] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 671.818929] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] yield resources [ 671.818929] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 671.818929] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] self.driver.spawn(context, instance, image_meta, [ 671.818929] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 671.818929] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 671.818929] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 671.818929] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] vm_ref = self.build_virtual_machine(instance, [ 671.818929] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 671.819312] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] vif_infos = vmwarevif.get_vif_info(self._session, [ 671.819312] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 671.819312] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] for vif in network_info: [ 671.819312] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 671.819312] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] return self._sync_wrapper(fn, *args, **kwargs) [ 671.819312] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 671.819312] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] self.wait() [ 671.819312] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 671.819312] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] self[:] = self._gt.wait() [ 671.819312] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 671.819312] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] return self._exit_event.wait() [ 671.819312] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 671.819312] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] result = hub.switch() [ 671.819719] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 671.819719] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] return self.greenlet.switch() [ 671.819719] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 671.819719] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] result = function(*args, **kwargs) [ 671.819719] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 671.819719] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] return func(*args, **kwargs) [ 671.819719] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 671.819719] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] raise e [ 671.819719] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 671.819719] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] nwinfo = self.network_api.allocate_for_instance( [ 671.819719] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 671.819719] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] created_port_ids = self._update_ports_for_instance( [ 671.819719] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 671.820125] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] with excutils.save_and_reraise_exception(): [ 671.820125] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.820125] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] self.force_reraise() [ 671.820125] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.820125] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] raise self.value [ 671.820125] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 671.820125] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] updated_port = self._update_port( [ 671.820125] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.820125] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] _ensure_no_port_binding_failure(port) [ 671.820125] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.820125] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] raise exception.PortBindingFailed(port_id=port['id']) [ 671.820125] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] nova.exception.PortBindingFailed: Binding failed for port 44e32753-d6cd-4e32-8b26-9476d40e482e, please check neutron logs for more information. [ 671.820125] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] [ 671.820587] env[62692]: INFO nova.compute.manager [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Terminating instance [ 671.822436] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Acquiring lock "refresh_cache-59882012-b01c-45ec-bd53-6e17e356ce5f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.172849] env[62692]: DEBUG nova.network.neutron [req-20824974-06ce-4110-a155-ea17f8b56ba1 req-f43bb296-665d-43f8-aaba-632c1d7cb177 service nova] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 672.231899] env[62692]: DEBUG nova.compute.utils [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 672.236519] env[62692]: DEBUG nova.compute.manager [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 672.238247] env[62692]: DEBUG nova.network.neutron [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 672.371195] env[62692]: DEBUG nova.policy [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b558d463509a40d697bc07a44e93bf55', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bac9418f752248caac5599268043abff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 672.381688] env[62692]: DEBUG nova.network.neutron [req-20824974-06ce-4110-a155-ea17f8b56ba1 req-f43bb296-665d-43f8-aaba-632c1d7cb177 service nova] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.737850] env[62692]: DEBUG nova.compute.manager [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 672.883179] env[62692]: DEBUG nova.network.neutron [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Successfully created port: c049b809-696e-4042-ada6-74917ad431b6 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 672.885015] env[62692]: DEBUG oslo_concurrency.lockutils [req-20824974-06ce-4110-a155-ea17f8b56ba1 req-f43bb296-665d-43f8-aaba-632c1d7cb177 service nova] Releasing lock "refresh_cache-59882012-b01c-45ec-bd53-6e17e356ce5f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.887878] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Acquired lock "refresh_cache-59882012-b01c-45ec-bd53-6e17e356ce5f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.888709] env[62692]: DEBUG nova.network.neutron [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 673.255539] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f66ac199-b508-46f5-8af9-15c3ff5cfbed {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.264551] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89424ee0-c019-4017-8b0f-e86c15cf0cec {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.300684] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9dd24e0-4427-4b69-b8fe-35f497eb72b0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.309049] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-745b1636-a425-4d6e-988c-f60efc2b73ab {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.324016] env[62692]: DEBUG nova.compute.provider_tree [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 673.413878] env[62692]: DEBUG nova.network.neutron [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 673.447506] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Acquiring lock "aa0666ce-4d4f-4333-8acb-ea0cfb438cd9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.448082] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Lock "aa0666ce-4d4f-4333-8acb-ea0cfb438cd9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.567021] env[62692]: DEBUG nova.network.neutron [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.757193] env[62692]: DEBUG nova.compute.manager [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 673.763431] env[62692]: DEBUG nova.compute.manager [req-487c5147-b66b-434d-9224-ead980d2d1c2 req-ff771538-e61f-4eed-b18e-c4c4f42b9b0e service nova] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Received event network-vif-deleted-44e32753-d6cd-4e32-8b26-9476d40e482e {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 673.792347] env[62692]: DEBUG nova.virt.hardware [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 673.792602] env[62692]: DEBUG nova.virt.hardware [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 673.792759] env[62692]: DEBUG nova.virt.hardware [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 673.792939] env[62692]: DEBUG nova.virt.hardware [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 673.793276] env[62692]: DEBUG nova.virt.hardware [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 673.793462] env[62692]: DEBUG nova.virt.hardware [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 673.793678] env[62692]: DEBUG nova.virt.hardware [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 673.793836] env[62692]: DEBUG nova.virt.hardware [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 673.794008] env[62692]: DEBUG nova.virt.hardware [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 673.794184] env[62692]: DEBUG nova.virt.hardware [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 673.794357] env[62692]: DEBUG nova.virt.hardware [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 673.795239] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8652cc3f-ae1e-4f7f-903a-1ce52d2d9256 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.805537] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e69b44c6-051b-42f8-91a8-73a981aa4354 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.827029] env[62692]: DEBUG nova.scheduler.client.report [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 674.066294] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Releasing lock "refresh_cache-59882012-b01c-45ec-bd53-6e17e356ce5f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.066761] env[62692]: DEBUG nova.compute.manager [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 674.066955] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 674.067279] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f43e8db6-f30a-43b9-a4a8-92ca133b5653 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.077351] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f233683c-efe8-4414-b2ee-56cf6949bdac {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.098401] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 59882012-b01c-45ec-bd53-6e17e356ce5f could not be found. [ 674.098659] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 674.098844] env[62692]: INFO nova.compute.manager [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Took 0.03 seconds to destroy the instance on the hypervisor. [ 674.099096] env[62692]: DEBUG oslo.service.loopingcall [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 674.099940] env[62692]: ERROR nova.compute.manager [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c049b809-696e-4042-ada6-74917ad431b6, please check neutron logs for more information. [ 674.099940] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 674.099940] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.099940] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 674.099940] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 674.099940] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 674.099940] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 674.099940] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 674.099940] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.099940] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 674.099940] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.099940] env[62692]: ERROR nova.compute.manager raise self.value [ 674.099940] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 674.099940] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 674.099940] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.099940] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 674.100543] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.100543] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 674.100543] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c049b809-696e-4042-ada6-74917ad431b6, please check neutron logs for more information. [ 674.100543] env[62692]: ERROR nova.compute.manager [ 674.100543] env[62692]: Traceback (most recent call last): [ 674.100543] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 674.100543] env[62692]: listener.cb(fileno) [ 674.100543] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.100543] env[62692]: result = function(*args, **kwargs) [ 674.100543] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 674.100543] env[62692]: return func(*args, **kwargs) [ 674.100543] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 674.100543] env[62692]: raise e [ 674.100543] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.100543] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 674.100543] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 674.100543] env[62692]: created_port_ids = self._update_ports_for_instance( [ 674.100543] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 674.100543] env[62692]: with excutils.save_and_reraise_exception(): [ 674.100543] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.100543] env[62692]: self.force_reraise() [ 674.100543] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.100543] env[62692]: raise self.value [ 674.100543] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 674.100543] env[62692]: updated_port = self._update_port( [ 674.100543] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.100543] env[62692]: _ensure_no_port_binding_failure(port) [ 674.100543] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.100543] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 674.101711] env[62692]: nova.exception.PortBindingFailed: Binding failed for port c049b809-696e-4042-ada6-74917ad431b6, please check neutron logs for more information. [ 674.101711] env[62692]: Removing descriptor: 19 [ 674.101711] env[62692]: DEBUG nova.compute.manager [-] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 674.101711] env[62692]: DEBUG nova.network.neutron [-] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 674.102402] env[62692]: ERROR nova.compute.manager [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c049b809-696e-4042-ada6-74917ad431b6, please check neutron logs for more information. [ 674.102402] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Traceback (most recent call last): [ 674.102402] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 674.102402] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] yield resources [ 674.102402] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 674.102402] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] self.driver.spawn(context, instance, image_meta, [ 674.102402] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 674.102402] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 674.102402] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 674.102402] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] vm_ref = self.build_virtual_machine(instance, [ 674.102402] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 674.102865] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] vif_infos = vmwarevif.get_vif_info(self._session, [ 674.102865] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 674.102865] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] for vif in network_info: [ 674.102865] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 674.102865] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] return self._sync_wrapper(fn, *args, **kwargs) [ 674.102865] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 674.102865] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] self.wait() [ 674.102865] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 674.102865] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] self[:] = self._gt.wait() [ 674.102865] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 674.102865] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] return self._exit_event.wait() [ 674.102865] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 674.102865] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] result = hub.switch() [ 674.103359] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 674.103359] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] return self.greenlet.switch() [ 674.103359] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.103359] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] result = function(*args, **kwargs) [ 674.103359] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 674.103359] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] return func(*args, **kwargs) [ 674.103359] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 674.103359] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] raise e [ 674.103359] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.103359] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] nwinfo = self.network_api.allocate_for_instance( [ 674.103359] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 674.103359] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] created_port_ids = self._update_ports_for_instance( [ 674.103359] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 674.103821] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] with excutils.save_and_reraise_exception(): [ 674.103821] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.103821] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] self.force_reraise() [ 674.103821] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.103821] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] raise self.value [ 674.103821] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 674.103821] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] updated_port = self._update_port( [ 674.103821] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.103821] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] _ensure_no_port_binding_failure(port) [ 674.103821] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.103821] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] raise exception.PortBindingFailed(port_id=port['id']) [ 674.103821] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] nova.exception.PortBindingFailed: Binding failed for port c049b809-696e-4042-ada6-74917ad431b6, please check neutron logs for more information. [ 674.103821] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] [ 674.104271] env[62692]: INFO nova.compute.manager [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Terminating instance [ 674.104701] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Acquiring lock "refresh_cache-b014c45b-4d27-4ade-b823-0dd64dfea2e2" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.104860] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Acquired lock "refresh_cache-b014c45b-4d27-4ade-b823-0dd64dfea2e2" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.105446] env[62692]: DEBUG nova.network.neutron [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 674.121203] env[62692]: DEBUG nova.network.neutron [-] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 674.332033] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.611s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.332568] env[62692]: DEBUG nova.compute.manager [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 674.335602] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.941s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.338177] env[62692]: INFO nova.compute.claims [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 674.623798] env[62692]: DEBUG nova.network.neutron [-] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.637209] env[62692]: DEBUG nova.network.neutron [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 674.739259] env[62692]: DEBUG nova.network.neutron [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.842212] env[62692]: DEBUG nova.compute.utils [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 674.845173] env[62692]: DEBUG nova.compute.manager [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 674.845277] env[62692]: DEBUG nova.network.neutron [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 674.886210] env[62692]: DEBUG nova.policy [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4e2556e3a4c4e158d8b457e7f4d8b61', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a0731ea847704b138016ecea565fa904', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 675.126542] env[62692]: INFO nova.compute.manager [-] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Took 1.03 seconds to deallocate network for instance. [ 675.129612] env[62692]: DEBUG nova.compute.claims [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 675.129879] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.173029] env[62692]: DEBUG nova.network.neutron [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Successfully created port: 6c9a2c2d-decd-4dc2-b0af-1c77575dd5cd {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 675.243406] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Releasing lock "refresh_cache-b014c45b-4d27-4ade-b823-0dd64dfea2e2" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.243406] env[62692]: DEBUG nova.compute.manager [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 675.243406] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 675.243406] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d49d1751-5a9d-4013-a242-6022aa4e7af6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.251755] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f4b5531-c6d6-420b-a908-cd691f2e58c8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.276050] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b014c45b-4d27-4ade-b823-0dd64dfea2e2 could not be found. [ 675.276050] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 675.276050] env[62692]: INFO nova.compute.manager [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Took 0.03 seconds to destroy the instance on the hypervisor. [ 675.276050] env[62692]: DEBUG oslo.service.loopingcall [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 675.276050] env[62692]: DEBUG nova.compute.manager [-] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 675.276050] env[62692]: DEBUG nova.network.neutron [-] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 675.294158] env[62692]: DEBUG nova.network.neutron [-] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 675.351052] env[62692]: DEBUG nova.compute.manager [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 675.797075] env[62692]: DEBUG nova.network.neutron [-] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.902041] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2b8e5fd-af7c-4484-b61a-b0419c2dcd35 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.912354] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c768290-212e-4e7e-9718-115a6ed25f1d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.950764] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e817bc4e-a94e-4723-a0ca-98290b247b97 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.954349] env[62692]: DEBUG nova.compute.manager [req-1fb0dc71-d55d-433b-a367-db168314674f req-60e7c9de-b151-4aeb-ae79-2aea1f16cfdf service nova] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Received event network-changed-c049b809-696e-4042-ada6-74917ad431b6 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 675.954693] env[62692]: DEBUG nova.compute.manager [req-1fb0dc71-d55d-433b-a367-db168314674f req-60e7c9de-b151-4aeb-ae79-2aea1f16cfdf service nova] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Refreshing instance network info cache due to event network-changed-c049b809-696e-4042-ada6-74917ad431b6. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 675.955038] env[62692]: DEBUG oslo_concurrency.lockutils [req-1fb0dc71-d55d-433b-a367-db168314674f req-60e7c9de-b151-4aeb-ae79-2aea1f16cfdf service nova] Acquiring lock "refresh_cache-b014c45b-4d27-4ade-b823-0dd64dfea2e2" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.955285] env[62692]: DEBUG oslo_concurrency.lockutils [req-1fb0dc71-d55d-433b-a367-db168314674f req-60e7c9de-b151-4aeb-ae79-2aea1f16cfdf service nova] Acquired lock "refresh_cache-b014c45b-4d27-4ade-b823-0dd64dfea2e2" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.955576] env[62692]: DEBUG nova.network.neutron [req-1fb0dc71-d55d-433b-a367-db168314674f req-60e7c9de-b151-4aeb-ae79-2aea1f16cfdf service nova] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Refreshing network info cache for port c049b809-696e-4042-ada6-74917ad431b6 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 675.963341] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e754abd6-0e11-4ffa-826b-39ffc8354ad3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.980801] env[62692]: DEBUG nova.compute.provider_tree [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.238773] env[62692]: DEBUG oslo_concurrency.lockutils [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Acquiring lock "1848c251-0a0c-4c36-8bd0-ff37befde2a0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 676.239041] env[62692]: DEBUG oslo_concurrency.lockutils [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Lock "1848c251-0a0c-4c36-8bd0-ff37befde2a0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.302075] env[62692]: INFO nova.compute.manager [-] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Took 1.03 seconds to deallocate network for instance. [ 676.304357] env[62692]: DEBUG nova.compute.claims [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 676.304529] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 676.361252] env[62692]: DEBUG nova.compute.manager [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 676.387995] env[62692]: DEBUG nova.virt.hardware [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 676.387995] env[62692]: DEBUG nova.virt.hardware [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 676.388190] env[62692]: DEBUG nova.virt.hardware [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 676.388399] env[62692]: DEBUG nova.virt.hardware [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 676.388674] env[62692]: DEBUG nova.virt.hardware [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 676.388878] env[62692]: DEBUG nova.virt.hardware [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 676.389143] env[62692]: DEBUG nova.virt.hardware [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 676.389344] env[62692]: DEBUG nova.virt.hardware [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 676.389556] env[62692]: DEBUG nova.virt.hardware [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 676.389763] env[62692]: DEBUG nova.virt.hardware [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 676.390015] env[62692]: DEBUG nova.virt.hardware [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 676.390902] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4f287d8-d97a-441a-8932-a975cf163ee4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.399709] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cf3547f-7294-473d-b910-d0d081e7e5cd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.483482] env[62692]: DEBUG nova.network.neutron [req-1fb0dc71-d55d-433b-a367-db168314674f req-60e7c9de-b151-4aeb-ae79-2aea1f16cfdf service nova] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 676.487900] env[62692]: DEBUG nova.scheduler.client.report [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 676.596024] env[62692]: ERROR nova.compute.manager [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6c9a2c2d-decd-4dc2-b0af-1c77575dd5cd, please check neutron logs for more information. [ 676.596024] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 676.596024] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 676.596024] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 676.596024] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 676.596024] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 676.596024] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 676.596024] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 676.596024] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.596024] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 676.596024] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.596024] env[62692]: ERROR nova.compute.manager raise self.value [ 676.596024] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 676.596024] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 676.596024] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 676.596024] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 676.596616] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 676.596616] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 676.596616] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6c9a2c2d-decd-4dc2-b0af-1c77575dd5cd, please check neutron logs for more information. [ 676.596616] env[62692]: ERROR nova.compute.manager [ 676.596616] env[62692]: Traceback (most recent call last): [ 676.596616] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 676.596616] env[62692]: listener.cb(fileno) [ 676.596616] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 676.596616] env[62692]: result = function(*args, **kwargs) [ 676.596616] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 676.596616] env[62692]: return func(*args, **kwargs) [ 676.596616] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 676.596616] env[62692]: raise e [ 676.596616] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 676.596616] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 676.596616] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 676.596616] env[62692]: created_port_ids = self._update_ports_for_instance( [ 676.596616] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 676.596616] env[62692]: with excutils.save_and_reraise_exception(): [ 676.596616] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.596616] env[62692]: self.force_reraise() [ 676.596616] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.596616] env[62692]: raise self.value [ 676.596616] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 676.596616] env[62692]: updated_port = self._update_port( [ 676.596616] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 676.596616] env[62692]: _ensure_no_port_binding_failure(port) [ 676.596616] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 676.596616] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 676.597795] env[62692]: nova.exception.PortBindingFailed: Binding failed for port 6c9a2c2d-decd-4dc2-b0af-1c77575dd5cd, please check neutron logs for more information. [ 676.597795] env[62692]: Removing descriptor: 19 [ 676.597795] env[62692]: ERROR nova.compute.manager [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6c9a2c2d-decd-4dc2-b0af-1c77575dd5cd, please check neutron logs for more information. [ 676.597795] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Traceback (most recent call last): [ 676.597795] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 676.597795] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] yield resources [ 676.597795] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 676.597795] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] self.driver.spawn(context, instance, image_meta, [ 676.597795] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 676.597795] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 676.597795] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 676.597795] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] vm_ref = self.build_virtual_machine(instance, [ 676.598221] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 676.598221] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] vif_infos = vmwarevif.get_vif_info(self._session, [ 676.598221] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 676.598221] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] for vif in network_info: [ 676.598221] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 676.598221] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] return self._sync_wrapper(fn, *args, **kwargs) [ 676.598221] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 676.598221] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] self.wait() [ 676.598221] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 676.598221] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] self[:] = self._gt.wait() [ 676.598221] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 676.598221] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] return self._exit_event.wait() [ 676.598221] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 676.598718] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] result = hub.switch() [ 676.598718] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 676.598718] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] return self.greenlet.switch() [ 676.598718] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 676.598718] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] result = function(*args, **kwargs) [ 676.598718] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 676.598718] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] return func(*args, **kwargs) [ 676.598718] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 676.598718] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] raise e [ 676.598718] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 676.598718] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] nwinfo = self.network_api.allocate_for_instance( [ 676.598718] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 676.598718] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] created_port_ids = self._update_ports_for_instance( [ 676.599179] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 676.599179] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] with excutils.save_and_reraise_exception(): [ 676.599179] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.599179] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] self.force_reraise() [ 676.599179] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.599179] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] raise self.value [ 676.599179] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 676.599179] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] updated_port = self._update_port( [ 676.599179] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 676.599179] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] _ensure_no_port_binding_failure(port) [ 676.599179] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 676.599179] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] raise exception.PortBindingFailed(port_id=port['id']) [ 676.599572] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] nova.exception.PortBindingFailed: Binding failed for port 6c9a2c2d-decd-4dc2-b0af-1c77575dd5cd, please check neutron logs for more information. [ 676.599572] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] [ 676.599572] env[62692]: INFO nova.compute.manager [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Terminating instance [ 676.599572] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Acquiring lock "refresh_cache-6883c19f-1e3a-4eac-be74-8e725faee77a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.599572] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Acquired lock "refresh_cache-6883c19f-1e3a-4eac-be74-8e725faee77a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.599572] env[62692]: DEBUG nova.network.neutron [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 676.633886] env[62692]: DEBUG nova.network.neutron [req-1fb0dc71-d55d-433b-a367-db168314674f req-60e7c9de-b151-4aeb-ae79-2aea1f16cfdf service nova] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.995629] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.660s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.996196] env[62692]: DEBUG nova.compute.manager [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 676.998808] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.589s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.120290] env[62692]: DEBUG nova.network.neutron [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.137251] env[62692]: DEBUG oslo_concurrency.lockutils [req-1fb0dc71-d55d-433b-a367-db168314674f req-60e7c9de-b151-4aeb-ae79-2aea1f16cfdf service nova] Releasing lock "refresh_cache-b014c45b-4d27-4ade-b823-0dd64dfea2e2" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.137538] env[62692]: DEBUG nova.compute.manager [req-1fb0dc71-d55d-433b-a367-db168314674f req-60e7c9de-b151-4aeb-ae79-2aea1f16cfdf service nova] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Received event network-vif-deleted-c049b809-696e-4042-ada6-74917ad431b6 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 677.227890] env[62692]: DEBUG nova.network.neutron [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.360313] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Acquiring lock "81ce340c-fbef-4932-983c-595843530dbc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.360313] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Lock "81ce340c-fbef-4932-983c-595843530dbc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.503034] env[62692]: DEBUG nova.compute.utils [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 677.508202] env[62692]: DEBUG nova.compute.manager [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 677.508311] env[62692]: DEBUG nova.network.neutron [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 677.579913] env[62692]: DEBUG nova.policy [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '073816e369134d71b0e4722e63583c6a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6e17ea308d264b60be1ed2488653327f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 677.731323] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Releasing lock "refresh_cache-6883c19f-1e3a-4eac-be74-8e725faee77a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.731769] env[62692]: DEBUG nova.compute.manager [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 677.731968] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 677.732274] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6ef9e30e-56f9-4f79-9eca-57f774771d81 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.740892] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a08e0004-3676-43f1-991d-be996a81b960 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.765469] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6883c19f-1e3a-4eac-be74-8e725faee77a could not be found. [ 677.765779] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 677.766043] env[62692]: INFO nova.compute.manager [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Took 0.03 seconds to destroy the instance on the hypervisor. [ 677.766392] env[62692]: DEBUG oslo.service.loopingcall [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 677.766678] env[62692]: DEBUG nova.compute.manager [-] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 677.766805] env[62692]: DEBUG nova.network.neutron [-] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 677.789112] env[62692]: DEBUG nova.network.neutron [-] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.985233] env[62692]: DEBUG nova.compute.manager [req-f78db725-32aa-4f3a-90c3-279dbd59a358 req-8ed581f5-d870-42f2-bb74-3ab1823e7cc6 service nova] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Received event network-changed-6c9a2c2d-decd-4dc2-b0af-1c77575dd5cd {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 677.985459] env[62692]: DEBUG nova.compute.manager [req-f78db725-32aa-4f3a-90c3-279dbd59a358 req-8ed581f5-d870-42f2-bb74-3ab1823e7cc6 service nova] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Refreshing instance network info cache due to event network-changed-6c9a2c2d-decd-4dc2-b0af-1c77575dd5cd. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 677.985771] env[62692]: DEBUG oslo_concurrency.lockutils [req-f78db725-32aa-4f3a-90c3-279dbd59a358 req-8ed581f5-d870-42f2-bb74-3ab1823e7cc6 service nova] Acquiring lock "refresh_cache-6883c19f-1e3a-4eac-be74-8e725faee77a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.985839] env[62692]: DEBUG oslo_concurrency.lockutils [req-f78db725-32aa-4f3a-90c3-279dbd59a358 req-8ed581f5-d870-42f2-bb74-3ab1823e7cc6 service nova] Acquired lock "refresh_cache-6883c19f-1e3a-4eac-be74-8e725faee77a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.986047] env[62692]: DEBUG nova.network.neutron [req-f78db725-32aa-4f3a-90c3-279dbd59a358 req-8ed581f5-d870-42f2-bb74-3ab1823e7cc6 service nova] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Refreshing network info cache for port 6c9a2c2d-decd-4dc2-b0af-1c77575dd5cd {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 678.008606] env[62692]: DEBUG nova.compute.manager [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 678.019216] env[62692]: DEBUG nova.network.neutron [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Successfully created port: cd7a8a64-e3fd-4852-b34b-f93d8452dd9a {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 678.047929] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0087705-0b30-4edb-b591-801ba0b784a5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.057844] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a150a0-f39b-4b5a-8db9-cea0b4cbda59 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.094944] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e254639e-76bf-4c08-bf60-40924955ff52 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.102284] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-046dec4e-f658-47c8-aa09-427e08827e77 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.118267] env[62692]: DEBUG nova.compute.provider_tree [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 678.293309] env[62692]: DEBUG nova.network.neutron [-] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.510924] env[62692]: DEBUG nova.network.neutron [req-f78db725-32aa-4f3a-90c3-279dbd59a358 req-8ed581f5-d870-42f2-bb74-3ab1823e7cc6 service nova] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 678.621223] env[62692]: DEBUG nova.scheduler.client.report [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 678.661749] env[62692]: DEBUG nova.network.neutron [req-f78db725-32aa-4f3a-90c3-279dbd59a358 req-8ed581f5-d870-42f2-bb74-3ab1823e7cc6 service nova] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.798335] env[62692]: INFO nova.compute.manager [-] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Took 1.03 seconds to deallocate network for instance. [ 678.801086] env[62692]: DEBUG nova.compute.claims [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 678.801424] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.019153] env[62692]: DEBUG nova.compute.manager [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 679.044451] env[62692]: DEBUG nova.virt.hardware [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 679.044688] env[62692]: DEBUG nova.virt.hardware [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 679.044845] env[62692]: DEBUG nova.virt.hardware [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 679.045096] env[62692]: DEBUG nova.virt.hardware [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 679.045210] env[62692]: DEBUG nova.virt.hardware [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 679.045346] env[62692]: DEBUG nova.virt.hardware [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 679.045629] env[62692]: DEBUG nova.virt.hardware [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 679.045791] env[62692]: DEBUG nova.virt.hardware [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 679.045967] env[62692]: DEBUG nova.virt.hardware [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 679.046237] env[62692]: DEBUG nova.virt.hardware [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 679.046279] env[62692]: DEBUG nova.virt.hardware [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 679.047182] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a4b7e31-922c-4a10-801e-0a0ca7e2da8a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.056241] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65d5f60b-a204-4b4d-a2d0-d8a3a9a08765 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.129019] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.128s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.129019] env[62692]: ERROR nova.compute.manager [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d37419d2-13ae-4fd0-8880-7dd92681d221, please check neutron logs for more information. [ 679.129019] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Traceback (most recent call last): [ 679.129019] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 679.129019] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] self.driver.spawn(context, instance, image_meta, [ 679.129019] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 679.129019] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] self._vmops.spawn(context, instance, image_meta, injected_files, [ 679.129019] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 679.129019] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] vm_ref = self.build_virtual_machine(instance, [ 679.129456] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 679.129456] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] vif_infos = vmwarevif.get_vif_info(self._session, [ 679.129456] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 679.129456] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] for vif in network_info: [ 679.129456] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 679.129456] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] return self._sync_wrapper(fn, *args, **kwargs) [ 679.129456] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 679.129456] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] self.wait() [ 679.129456] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 679.129456] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] self[:] = self._gt.wait() [ 679.129456] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 679.129456] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] return self._exit_event.wait() [ 679.129456] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 679.129886] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] result = hub.switch() [ 679.129886] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 679.129886] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] return self.greenlet.switch() [ 679.129886] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.129886] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] result = function(*args, **kwargs) [ 679.129886] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 679.129886] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] return func(*args, **kwargs) [ 679.129886] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 679.129886] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] raise e [ 679.129886] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.129886] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] nwinfo = self.network_api.allocate_for_instance( [ 679.129886] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 679.129886] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] created_port_ids = self._update_ports_for_instance( [ 679.130306] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 679.130306] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] with excutils.save_and_reraise_exception(): [ 679.130306] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.130306] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] self.force_reraise() [ 679.130306] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.130306] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] raise self.value [ 679.130306] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 679.130306] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] updated_port = self._update_port( [ 679.130306] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.130306] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] _ensure_no_port_binding_failure(port) [ 679.130306] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.130306] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] raise exception.PortBindingFailed(port_id=port['id']) [ 679.130695] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] nova.exception.PortBindingFailed: Binding failed for port d37419d2-13ae-4fd0-8880-7dd92681d221, please check neutron logs for more information. [ 679.130695] env[62692]: ERROR nova.compute.manager [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] [ 679.130695] env[62692]: DEBUG nova.compute.utils [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Binding failed for port d37419d2-13ae-4fd0-8880-7dd92681d221, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 679.132265] env[62692]: ERROR nova.compute.manager [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cd7a8a64-e3fd-4852-b34b-f93d8452dd9a, please check neutron logs for more information. [ 679.132265] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 679.132265] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.132265] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 679.132265] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 679.132265] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 679.132265] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 679.132265] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 679.132265] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.132265] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 679.132265] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.132265] env[62692]: ERROR nova.compute.manager raise self.value [ 679.132265] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 679.132265] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 679.132265] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.132265] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 679.132814] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.132814] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 679.132814] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cd7a8a64-e3fd-4852-b34b-f93d8452dd9a, please check neutron logs for more information. [ 679.132814] env[62692]: ERROR nova.compute.manager [ 679.132814] env[62692]: Traceback (most recent call last): [ 679.132814] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 679.132814] env[62692]: listener.cb(fileno) [ 679.132814] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.132814] env[62692]: result = function(*args, **kwargs) [ 679.132814] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 679.132814] env[62692]: return func(*args, **kwargs) [ 679.132814] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 679.132814] env[62692]: raise e [ 679.132814] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.132814] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 679.132814] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 679.132814] env[62692]: created_port_ids = self._update_ports_for_instance( [ 679.132814] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 679.132814] env[62692]: with excutils.save_and_reraise_exception(): [ 679.132814] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.132814] env[62692]: self.force_reraise() [ 679.132814] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.132814] env[62692]: raise self.value [ 679.132814] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 679.132814] env[62692]: updated_port = self._update_port( [ 679.132814] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.132814] env[62692]: _ensure_no_port_binding_failure(port) [ 679.132814] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.132814] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 679.133728] env[62692]: nova.exception.PortBindingFailed: Binding failed for port cd7a8a64-e3fd-4852-b34b-f93d8452dd9a, please check neutron logs for more information. [ 679.133728] env[62692]: Removing descriptor: 17 [ 679.133728] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.474s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.134170] env[62692]: INFO nova.compute.claims [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 679.138110] env[62692]: ERROR nova.compute.manager [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cd7a8a64-e3fd-4852-b34b-f93d8452dd9a, please check neutron logs for more information. [ 679.138110] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] Traceback (most recent call last): [ 679.138110] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 679.138110] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] yield resources [ 679.138110] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 679.138110] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] self.driver.spawn(context, instance, image_meta, [ 679.138110] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 679.138110] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 679.138110] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 679.138110] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] vm_ref = self.build_virtual_machine(instance, [ 679.138110] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 679.138582] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] vif_infos = vmwarevif.get_vif_info(self._session, [ 679.138582] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 679.138582] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] for vif in network_info: [ 679.138582] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 679.138582] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] return self._sync_wrapper(fn, *args, **kwargs) [ 679.138582] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 679.138582] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] self.wait() [ 679.138582] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 679.138582] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] self[:] = self._gt.wait() [ 679.138582] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 679.138582] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] return self._exit_event.wait() [ 679.138582] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 679.138582] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] result = hub.switch() [ 679.138972] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 679.138972] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] return self.greenlet.switch() [ 679.138972] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.138972] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] result = function(*args, **kwargs) [ 679.138972] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 679.138972] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] return func(*args, **kwargs) [ 679.138972] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 679.138972] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] raise e [ 679.138972] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.138972] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] nwinfo = self.network_api.allocate_for_instance( [ 679.138972] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 679.138972] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] created_port_ids = self._update_ports_for_instance( [ 679.138972] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 679.139372] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] with excutils.save_and_reraise_exception(): [ 679.139372] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.139372] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] self.force_reraise() [ 679.139372] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.139372] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] raise self.value [ 679.139372] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 679.139372] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] updated_port = self._update_port( [ 679.139372] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.139372] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] _ensure_no_port_binding_failure(port) [ 679.139372] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.139372] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] raise exception.PortBindingFailed(port_id=port['id']) [ 679.139372] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] nova.exception.PortBindingFailed: Binding failed for port cd7a8a64-e3fd-4852-b34b-f93d8452dd9a, please check neutron logs for more information. [ 679.139372] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] [ 679.139809] env[62692]: INFO nova.compute.manager [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Terminating instance [ 679.140828] env[62692]: DEBUG nova.compute.manager [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Build of instance f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed was re-scheduled: Binding failed for port d37419d2-13ae-4fd0-8880-7dd92681d221, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 679.141266] env[62692]: DEBUG nova.compute.manager [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 679.141316] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Acquiring lock "refresh_cache-f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.141467] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Acquired lock "refresh_cache-f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.141628] env[62692]: DEBUG nova.network.neutron [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 679.150981] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Acquiring lock "refresh_cache-844b3803-8a01-484e-8908-dbae27188acb" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.151172] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Acquired lock "refresh_cache-844b3803-8a01-484e-8908-dbae27188acb" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.151346] env[62692]: DEBUG nova.network.neutron [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 679.164342] env[62692]: DEBUG oslo_concurrency.lockutils [req-f78db725-32aa-4f3a-90c3-279dbd59a358 req-8ed581f5-d870-42f2-bb74-3ab1823e7cc6 service nova] Releasing lock "refresh_cache-6883c19f-1e3a-4eac-be74-8e725faee77a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.164600] env[62692]: DEBUG nova.compute.manager [req-f78db725-32aa-4f3a-90c3-279dbd59a358 req-8ed581f5-d870-42f2-bb74-3ab1823e7cc6 service nova] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Received event network-vif-deleted-6c9a2c2d-decd-4dc2-b0af-1c77575dd5cd {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 679.662786] env[62692]: DEBUG nova.network.neutron [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 679.669973] env[62692]: DEBUG nova.network.neutron [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 679.755701] env[62692]: DEBUG nova.network.neutron [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.780248] env[62692]: DEBUG nova.network.neutron [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.032188] env[62692]: DEBUG nova.compute.manager [req-989d45f9-af0c-4ccb-be69-7e0dc9e949f2 req-7e403d7a-db19-4c32-a448-367ff23269e2 service nova] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Received event network-changed-cd7a8a64-e3fd-4852-b34b-f93d8452dd9a {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 680.032385] env[62692]: DEBUG nova.compute.manager [req-989d45f9-af0c-4ccb-be69-7e0dc9e949f2 req-7e403d7a-db19-4c32-a448-367ff23269e2 service nova] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Refreshing instance network info cache due to event network-changed-cd7a8a64-e3fd-4852-b34b-f93d8452dd9a. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 680.032619] env[62692]: DEBUG oslo_concurrency.lockutils [req-989d45f9-af0c-4ccb-be69-7e0dc9e949f2 req-7e403d7a-db19-4c32-a448-367ff23269e2 service nova] Acquiring lock "refresh_cache-844b3803-8a01-484e-8908-dbae27188acb" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.261090] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Releasing lock "refresh_cache-844b3803-8a01-484e-8908-dbae27188acb" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.261090] env[62692]: DEBUG nova.compute.manager [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 680.261485] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 680.262109] env[62692]: DEBUG oslo_concurrency.lockutils [req-989d45f9-af0c-4ccb-be69-7e0dc9e949f2 req-7e403d7a-db19-4c32-a448-367ff23269e2 service nova] Acquired lock "refresh_cache-844b3803-8a01-484e-8908-dbae27188acb" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.263734] env[62692]: DEBUG nova.network.neutron [req-989d45f9-af0c-4ccb-be69-7e0dc9e949f2 req-7e403d7a-db19-4c32-a448-367ff23269e2 service nova] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Refreshing network info cache for port cd7a8a64-e3fd-4852-b34b-f93d8452dd9a {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 680.263734] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0c5c3843-2696-4b5b-991a-e3ef1602512f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.280176] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12f734e7-04d1-445e-a62a-20dd24148e91 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.291550] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Releasing lock "refresh_cache-f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.292382] env[62692]: DEBUG nova.compute.manager [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 680.292382] env[62692]: DEBUG nova.compute.manager [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 680.292382] env[62692]: DEBUG nova.network.neutron [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 680.307035] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 844b3803-8a01-484e-8908-dbae27188acb could not be found. [ 680.307035] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 680.307193] env[62692]: INFO nova.compute.manager [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Took 0.05 seconds to destroy the instance on the hypervisor. [ 680.309107] env[62692]: DEBUG oslo.service.loopingcall [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 680.309107] env[62692]: DEBUG nova.compute.manager [-] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 680.309107] env[62692]: DEBUG nova.network.neutron [-] [instance: 844b3803-8a01-484e-8908-dbae27188acb] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 680.317330] env[62692]: DEBUG nova.network.neutron [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 680.339683] env[62692]: DEBUG nova.network.neutron [-] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 680.668372] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c01798f-b8a9-4446-8091-679d01648c36 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.676531] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90d84432-0292-4e82-8fab-ca9e9dc3a7dd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.706974] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f76b9bf-fb45-4692-b34e-97ab2132dc2f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.714101] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db9cbe07-c2c7-41af-85ad-a5fd9a91c7b4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.726582] env[62692]: DEBUG nova.compute.provider_tree [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 680.782994] env[62692]: DEBUG nova.network.neutron [req-989d45f9-af0c-4ccb-be69-7e0dc9e949f2 req-7e403d7a-db19-4c32-a448-367ff23269e2 service nova] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 680.819797] env[62692]: DEBUG nova.network.neutron [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.826146] env[62692]: DEBUG nova.network.neutron [req-989d45f9-af0c-4ccb-be69-7e0dc9e949f2 req-7e403d7a-db19-4c32-a448-367ff23269e2 service nova] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.843022] env[62692]: DEBUG nova.network.neutron [-] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.232078] env[62692]: DEBUG nova.scheduler.client.report [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 681.322102] env[62692]: INFO nova.compute.manager [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] [instance: f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed] Took 1.03 seconds to deallocate network for instance. [ 681.331329] env[62692]: DEBUG oslo_concurrency.lockutils [req-989d45f9-af0c-4ccb-be69-7e0dc9e949f2 req-7e403d7a-db19-4c32-a448-367ff23269e2 service nova] Releasing lock "refresh_cache-844b3803-8a01-484e-8908-dbae27188acb" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.331562] env[62692]: DEBUG nova.compute.manager [req-989d45f9-af0c-4ccb-be69-7e0dc9e949f2 req-7e403d7a-db19-4c32-a448-367ff23269e2 service nova] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Received event network-vif-deleted-cd7a8a64-e3fd-4852-b34b-f93d8452dd9a {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 681.345168] env[62692]: INFO nova.compute.manager [-] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Took 1.04 seconds to deallocate network for instance. [ 681.347573] env[62692]: DEBUG nova.compute.claims [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 681.347642] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 681.736984] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.604s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.737615] env[62692]: DEBUG nova.compute.manager [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 681.740368] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.426s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 681.741751] env[62692]: INFO nova.compute.claims [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 682.246663] env[62692]: DEBUG nova.compute.utils [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 682.250667] env[62692]: DEBUG nova.compute.manager [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 682.250667] env[62692]: DEBUG nova.network.neutron [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 682.312404] env[62692]: DEBUG nova.policy [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '073816e369134d71b0e4722e63583c6a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6e17ea308d264b60be1ed2488653327f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 682.356190] env[62692]: INFO nova.scheduler.client.report [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Deleted allocations for instance f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed [ 682.680795] env[62692]: DEBUG nova.network.neutron [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Successfully created port: 32670d72-1490-4454-aa34-853d890b202b {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 682.755057] env[62692]: DEBUG nova.compute.manager [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 682.864475] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7de2d478-42b4-4b40-af56-02b4ef9b1cb0 tempest-ListServerFiltersTestJSON-862780760 tempest-ListServerFiltersTestJSON-862780760-project-member] Lock "f0a636a9-68c6-4b90-9dd3-3c0a0da4a0ed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.891s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.332248] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0b35b4f-8ef0-425a-a149-2e7d9a6c4611 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.340932] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-266fc4dd-6fca-453a-a0f5-6f1084222c09 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.374245] env[62692]: DEBUG nova.compute.manager [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 683.377458] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c91eecb-f84f-47f1-87bb-ad774dc06b05 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.385487] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6f4a91b-6b10-45fe-936e-1563ab95ada0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.399736] env[62692]: DEBUG nova.compute.provider_tree [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.702994] env[62692]: DEBUG nova.compute.manager [req-418f320f-127a-4524-85f3-5d355cb460ef req-c811c227-b7b7-4a4a-921f-a8375aa1c30b service nova] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Received event network-changed-32670d72-1490-4454-aa34-853d890b202b {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 683.703111] env[62692]: DEBUG nova.compute.manager [req-418f320f-127a-4524-85f3-5d355cb460ef req-c811c227-b7b7-4a4a-921f-a8375aa1c30b service nova] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Refreshing instance network info cache due to event network-changed-32670d72-1490-4454-aa34-853d890b202b. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 683.703358] env[62692]: DEBUG oslo_concurrency.lockutils [req-418f320f-127a-4524-85f3-5d355cb460ef req-c811c227-b7b7-4a4a-921f-a8375aa1c30b service nova] Acquiring lock "refresh_cache-0aac5bf9-2af4-4745-a961-46986fb57779" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.703465] env[62692]: DEBUG oslo_concurrency.lockutils [req-418f320f-127a-4524-85f3-5d355cb460ef req-c811c227-b7b7-4a4a-921f-a8375aa1c30b service nova] Acquired lock "refresh_cache-0aac5bf9-2af4-4745-a961-46986fb57779" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.703633] env[62692]: DEBUG nova.network.neutron [req-418f320f-127a-4524-85f3-5d355cb460ef req-c811c227-b7b7-4a4a-921f-a8375aa1c30b service nova] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Refreshing network info cache for port 32670d72-1490-4454-aa34-853d890b202b {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 683.768726] env[62692]: DEBUG nova.compute.manager [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 683.806193] env[62692]: DEBUG nova.virt.hardware [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 683.806193] env[62692]: DEBUG nova.virt.hardware [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 683.806193] env[62692]: DEBUG nova.virt.hardware [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 683.810026] env[62692]: DEBUG nova.virt.hardware [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 683.810026] env[62692]: DEBUG nova.virt.hardware [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 683.810026] env[62692]: DEBUG nova.virt.hardware [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 683.810026] env[62692]: DEBUG nova.virt.hardware [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 683.810026] env[62692]: DEBUG nova.virt.hardware [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 683.810316] env[62692]: DEBUG nova.virt.hardware [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 683.810316] env[62692]: DEBUG nova.virt.hardware [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 683.810316] env[62692]: DEBUG nova.virt.hardware [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 683.810316] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f6cdd9-0673-4493-b281-c96245e36aff {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.812626] env[62692]: ERROR nova.compute.manager [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 32670d72-1490-4454-aa34-853d890b202b, please check neutron logs for more information. [ 683.812626] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 683.812626] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 683.812626] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 683.812626] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 683.812626] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 683.812626] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 683.812626] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 683.812626] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 683.812626] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 683.812626] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 683.812626] env[62692]: ERROR nova.compute.manager raise self.value [ 683.812626] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 683.812626] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 683.812626] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 683.812626] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 683.813048] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 683.813048] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 683.813048] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 32670d72-1490-4454-aa34-853d890b202b, please check neutron logs for more information. [ 683.813048] env[62692]: ERROR nova.compute.manager [ 683.813390] env[62692]: Traceback (most recent call last): [ 683.813506] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 683.813506] env[62692]: listener.cb(fileno) [ 683.813610] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 683.813610] env[62692]: result = function(*args, **kwargs) [ 683.813671] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 683.813671] env[62692]: return func(*args, **kwargs) [ 683.813731] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 683.813731] env[62692]: raise e [ 683.813796] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 683.813796] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 683.813856] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 683.813856] env[62692]: created_port_ids = self._update_ports_for_instance( [ 683.813958] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 683.813958] env[62692]: with excutils.save_and_reraise_exception(): [ 683.814035] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 683.814035] env[62692]: self.force_reraise() [ 683.814104] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 683.814104] env[62692]: raise self.value [ 683.814163] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 683.814163] env[62692]: updated_port = self._update_port( [ 683.814221] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 683.814221] env[62692]: _ensure_no_port_binding_failure(port) [ 683.814278] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 683.814278] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 683.814455] env[62692]: nova.exception.PortBindingFailed: Binding failed for port 32670d72-1490-4454-aa34-853d890b202b, please check neutron logs for more information. [ 683.814500] env[62692]: Removing descriptor: 19 [ 683.820838] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39c07a2d-b2cf-4e31-86e4-5ce865bdab6d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.840216] env[62692]: ERROR nova.compute.manager [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 32670d72-1490-4454-aa34-853d890b202b, please check neutron logs for more information. [ 683.840216] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Traceback (most recent call last): [ 683.840216] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 683.840216] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] yield resources [ 683.840216] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 683.840216] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] self.driver.spawn(context, instance, image_meta, [ 683.840216] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 683.840216] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] self._vmops.spawn(context, instance, image_meta, injected_files, [ 683.840216] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 683.840216] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] vm_ref = self.build_virtual_machine(instance, [ 683.840216] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 683.840526] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] vif_infos = vmwarevif.get_vif_info(self._session, [ 683.840526] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 683.840526] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] for vif in network_info: [ 683.840526] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 683.840526] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] return self._sync_wrapper(fn, *args, **kwargs) [ 683.840526] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 683.840526] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] self.wait() [ 683.840526] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 683.840526] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] self[:] = self._gt.wait() [ 683.840526] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 683.840526] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] return self._exit_event.wait() [ 683.840526] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 683.840526] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] current.throw(*self._exc) [ 683.840816] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 683.840816] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] result = function(*args, **kwargs) [ 683.840816] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 683.840816] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] return func(*args, **kwargs) [ 683.840816] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 683.840816] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] raise e [ 683.840816] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 683.840816] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] nwinfo = self.network_api.allocate_for_instance( [ 683.840816] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 683.840816] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] created_port_ids = self._update_ports_for_instance( [ 683.840816] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 683.840816] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] with excutils.save_and_reraise_exception(): [ 683.840816] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 683.841113] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] self.force_reraise() [ 683.841113] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 683.841113] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] raise self.value [ 683.841113] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 683.841113] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] updated_port = self._update_port( [ 683.841113] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 683.841113] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] _ensure_no_port_binding_failure(port) [ 683.841113] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 683.841113] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] raise exception.PortBindingFailed(port_id=port['id']) [ 683.841113] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] nova.exception.PortBindingFailed: Binding failed for port 32670d72-1490-4454-aa34-853d890b202b, please check neutron logs for more information. [ 683.841113] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] [ 683.841113] env[62692]: INFO nova.compute.manager [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Terminating instance [ 683.842847] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Acquiring lock "refresh_cache-0aac5bf9-2af4-4745-a961-46986fb57779" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.902054] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.902943] env[62692]: DEBUG nova.scheduler.client.report [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 684.222051] env[62692]: DEBUG nova.network.neutron [req-418f320f-127a-4524-85f3-5d355cb460ef req-c811c227-b7b7-4a4a-921f-a8375aa1c30b service nova] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 684.288069] env[62692]: DEBUG nova.network.neutron [req-418f320f-127a-4524-85f3-5d355cb460ef req-c811c227-b7b7-4a4a-921f-a8375aa1c30b service nova] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.408332] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.668s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.408873] env[62692]: DEBUG nova.compute.manager [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 684.411972] env[62692]: DEBUG oslo_concurrency.lockutils [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.811s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.412200] env[62692]: DEBUG nova.objects.instance [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Lazy-loading 'resources' on Instance uuid 54f615e3-5c6e-44bb-aeb7-5620a639b55c {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 684.792366] env[62692]: DEBUG oslo_concurrency.lockutils [req-418f320f-127a-4524-85f3-5d355cb460ef req-c811c227-b7b7-4a4a-921f-a8375aa1c30b service nova] Releasing lock "refresh_cache-0aac5bf9-2af4-4745-a961-46986fb57779" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.792791] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Acquired lock "refresh_cache-0aac5bf9-2af4-4745-a961-46986fb57779" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.793087] env[62692]: DEBUG nova.network.neutron [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 684.915494] env[62692]: DEBUG nova.compute.utils [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 684.920040] env[62692]: DEBUG nova.compute.manager [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 684.920221] env[62692]: DEBUG nova.network.neutron [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 685.006844] env[62692]: DEBUG nova.policy [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '37efc111d7184ed7a2a91e4ff01c7224', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b958c528d4354f9086543e9f3a339744', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 685.421191] env[62692]: DEBUG nova.compute.manager [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 685.429166] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c124d18e-9896-407e-8d73-2e2c24492168 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.434501] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-709ac193-2192-4d65-81a2-f643494ec41f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.470987] env[62692]: DEBUG nova.network.neutron [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 685.473258] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46bad8e7-bcd3-4885-8abf-40a250a0e602 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.482117] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a105a29-c3f7-4ee4-8b36-a26d2233bae9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.508219] env[62692]: DEBUG nova.compute.provider_tree [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 685.581957] env[62692]: DEBUG nova.network.neutron [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.617725] env[62692]: DEBUG nova.network.neutron [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Successfully created port: 501c0ab4-2706-44d5-9057-d5ae4b34ae1f {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 685.826569] env[62692]: DEBUG nova.compute.manager [req-a20f7c27-ca15-4fe5-868f-4ae8103f42a7 req-b47e9c29-f511-4706-a739-f71aa1db44cb service nova] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Received event network-vif-deleted-32670d72-1490-4454-aa34-853d890b202b {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 686.014309] env[62692]: DEBUG nova.scheduler.client.report [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 686.086298] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Releasing lock "refresh_cache-0aac5bf9-2af4-4745-a961-46986fb57779" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.086771] env[62692]: DEBUG nova.compute.manager [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 686.086903] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 686.089171] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-28550c42-a38e-4e69-a541-1b973373adf0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.103881] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b679edd-9d21-4b76-ac27-739484c39503 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.130371] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0aac5bf9-2af4-4745-a961-46986fb57779 could not be found. [ 686.130635] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 686.130815] env[62692]: INFO nova.compute.manager [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Took 0.04 seconds to destroy the instance on the hypervisor. [ 686.131074] env[62692]: DEBUG oslo.service.loopingcall [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 686.131283] env[62692]: DEBUG nova.compute.manager [-] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 686.131374] env[62692]: DEBUG nova.network.neutron [-] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 686.150614] env[62692]: DEBUG nova.network.neutron [-] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 686.437386] env[62692]: DEBUG nova.compute.manager [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 686.473033] env[62692]: DEBUG nova.virt.hardware [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 686.473033] env[62692]: DEBUG nova.virt.hardware [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 686.473033] env[62692]: DEBUG nova.virt.hardware [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 686.473401] env[62692]: DEBUG nova.virt.hardware [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 686.473692] env[62692]: DEBUG nova.virt.hardware [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 686.476571] env[62692]: DEBUG nova.virt.hardware [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 686.476571] env[62692]: DEBUG nova.virt.hardware [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 686.476571] env[62692]: DEBUG nova.virt.hardware [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 686.476571] env[62692]: DEBUG nova.virt.hardware [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 686.476571] env[62692]: DEBUG nova.virt.hardware [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 686.476915] env[62692]: DEBUG nova.virt.hardware [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 686.476915] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dc11ac5-bfce-4e5a-951f-0a99a5dd2cec {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.489123] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-858cbbdc-2fee-477b-a9a0-329981c11261 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.522019] env[62692]: DEBUG oslo_concurrency.lockutils [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.108s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.522636] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.829s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.548013] env[62692]: INFO nova.scheduler.client.report [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Deleted allocations for instance 54f615e3-5c6e-44bb-aeb7-5620a639b55c [ 686.653396] env[62692]: DEBUG nova.network.neutron [-] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.732385] env[62692]: ERROR nova.compute.manager [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 501c0ab4-2706-44d5-9057-d5ae4b34ae1f, please check neutron logs for more information. [ 686.732385] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 686.732385] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 686.732385] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 686.732385] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 686.732385] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 686.732385] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 686.732385] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 686.732385] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.732385] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 686.732385] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.732385] env[62692]: ERROR nova.compute.manager raise self.value [ 686.732385] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 686.732385] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 686.732385] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.732385] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 686.732828] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.732828] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 686.732828] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 501c0ab4-2706-44d5-9057-d5ae4b34ae1f, please check neutron logs for more information. [ 686.732828] env[62692]: ERROR nova.compute.manager [ 686.732968] env[62692]: Traceback (most recent call last): [ 686.732968] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 686.732968] env[62692]: listener.cb(fileno) [ 686.732968] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 686.732968] env[62692]: result = function(*args, **kwargs) [ 686.732968] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 686.732968] env[62692]: return func(*args, **kwargs) [ 686.732968] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 686.732968] env[62692]: raise e [ 686.732968] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 686.732968] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 686.732968] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 686.732968] env[62692]: created_port_ids = self._update_ports_for_instance( [ 686.732968] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 686.732968] env[62692]: with excutils.save_and_reraise_exception(): [ 686.732968] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.732968] env[62692]: self.force_reraise() [ 686.732968] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.732968] env[62692]: raise self.value [ 686.732968] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 686.732968] env[62692]: updated_port = self._update_port( [ 686.732968] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.732968] env[62692]: _ensure_no_port_binding_failure(port) [ 686.732968] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.732968] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 686.733590] env[62692]: nova.exception.PortBindingFailed: Binding failed for port 501c0ab4-2706-44d5-9057-d5ae4b34ae1f, please check neutron logs for more information. [ 686.733590] env[62692]: Removing descriptor: 19 [ 686.733755] env[62692]: ERROR nova.compute.manager [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 501c0ab4-2706-44d5-9057-d5ae4b34ae1f, please check neutron logs for more information. [ 686.733755] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Traceback (most recent call last): [ 686.733755] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 686.733755] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] yield resources [ 686.733755] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 686.733755] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] self.driver.spawn(context, instance, image_meta, [ 686.733755] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 686.733755] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 686.733755] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 686.733755] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] vm_ref = self.build_virtual_machine(instance, [ 686.733755] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 686.734107] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] vif_infos = vmwarevif.get_vif_info(self._session, [ 686.734107] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 686.734107] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] for vif in network_info: [ 686.734107] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 686.734107] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] return self._sync_wrapper(fn, *args, **kwargs) [ 686.734107] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 686.734107] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] self.wait() [ 686.734107] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 686.734107] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] self[:] = self._gt.wait() [ 686.734107] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 686.734107] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] return self._exit_event.wait() [ 686.734107] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 686.734107] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] result = hub.switch() [ 686.734564] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 686.734564] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] return self.greenlet.switch() [ 686.734564] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 686.734564] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] result = function(*args, **kwargs) [ 686.734564] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 686.734564] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] return func(*args, **kwargs) [ 686.734564] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 686.734564] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] raise e [ 686.734564] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 686.734564] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] nwinfo = self.network_api.allocate_for_instance( [ 686.734564] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 686.734564] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] created_port_ids = self._update_ports_for_instance( [ 686.734564] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 686.734892] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] with excutils.save_and_reraise_exception(): [ 686.734892] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.734892] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] self.force_reraise() [ 686.734892] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.734892] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] raise self.value [ 686.734892] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 686.734892] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] updated_port = self._update_port( [ 686.734892] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.734892] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] _ensure_no_port_binding_failure(port) [ 686.734892] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.734892] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] raise exception.PortBindingFailed(port_id=port['id']) [ 686.734892] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] nova.exception.PortBindingFailed: Binding failed for port 501c0ab4-2706-44d5-9057-d5ae4b34ae1f, please check neutron logs for more information. [ 686.734892] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] [ 686.735232] env[62692]: INFO nova.compute.manager [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Terminating instance [ 686.739017] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Acquiring lock "refresh_cache-5d18922d-5937-47c8-a04a-aa757c153e8f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.743020] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Acquired lock "refresh_cache-5d18922d-5937-47c8-a04a-aa757c153e8f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.743020] env[62692]: DEBUG nova.network.neutron [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 687.057303] env[62692]: DEBUG oslo_concurrency.lockutils [None req-764b2e9f-34e1-480d-9a43-afb532c40c1b tempest-ServerDiagnosticsV248Test-1906508807 tempest-ServerDiagnosticsV248Test-1906508807-project-member] Lock "54f615e3-5c6e-44bb-aeb7-5620a639b55c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.277s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.156863] env[62692]: INFO nova.compute.manager [-] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Took 1.02 seconds to deallocate network for instance. [ 687.161887] env[62692]: DEBUG nova.compute.claims [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 687.162110] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.264309] env[62692]: DEBUG nova.network.neutron [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 687.370278] env[62692]: DEBUG nova.network.neutron [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.530646] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8db827b-e273-4529-8e32-95bc76b17d2e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.539543] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0454d88-9fe2-4176-98a7-42f17affb2f8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.570756] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caeb32a1-9555-4bbb-93a8-d482164f2ae9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.578498] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59b96c55-7eab-45e6-9acc-e63305b9a603 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.592252] env[62692]: DEBUG nova.compute.provider_tree [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 687.873846] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Releasing lock "refresh_cache-5d18922d-5937-47c8-a04a-aa757c153e8f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.873846] env[62692]: DEBUG nova.compute.manager [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 687.873846] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 687.873846] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-94668c5e-2bcd-4779-968e-9a18aac0f61f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.884171] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-629d1c97-cf59-4551-9c4c-cd9ad1fdc814 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.896181] env[62692]: DEBUG nova.compute.manager [req-491d5e80-a905-429d-bee8-6747479761c9 req-d3eaeb11-7f71-4c35-8bcc-76533385355a service nova] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Received event network-changed-501c0ab4-2706-44d5-9057-d5ae4b34ae1f {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 687.896375] env[62692]: DEBUG nova.compute.manager [req-491d5e80-a905-429d-bee8-6747479761c9 req-d3eaeb11-7f71-4c35-8bcc-76533385355a service nova] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Refreshing instance network info cache due to event network-changed-501c0ab4-2706-44d5-9057-d5ae4b34ae1f. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 687.896583] env[62692]: DEBUG oslo_concurrency.lockutils [req-491d5e80-a905-429d-bee8-6747479761c9 req-d3eaeb11-7f71-4c35-8bcc-76533385355a service nova] Acquiring lock "refresh_cache-5d18922d-5937-47c8-a04a-aa757c153e8f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.896729] env[62692]: DEBUG oslo_concurrency.lockutils [req-491d5e80-a905-429d-bee8-6747479761c9 req-d3eaeb11-7f71-4c35-8bcc-76533385355a service nova] Acquired lock "refresh_cache-5d18922d-5937-47c8-a04a-aa757c153e8f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.896888] env[62692]: DEBUG nova.network.neutron [req-491d5e80-a905-429d-bee8-6747479761c9 req-d3eaeb11-7f71-4c35-8bcc-76533385355a service nova] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Refreshing network info cache for port 501c0ab4-2706-44d5-9057-d5ae4b34ae1f {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 687.909681] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5d18922d-5937-47c8-a04a-aa757c153e8f could not be found. [ 687.909756] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 687.909932] env[62692]: INFO nova.compute.manager [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 687.910420] env[62692]: DEBUG oslo.service.loopingcall [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 687.910908] env[62692]: DEBUG nova.compute.manager [-] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 687.911675] env[62692]: DEBUG nova.network.neutron [-] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 687.951673] env[62692]: DEBUG nova.network.neutron [-] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 688.096479] env[62692]: DEBUG nova.scheduler.client.report [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 688.433819] env[62692]: DEBUG nova.network.neutron [req-491d5e80-a905-429d-bee8-6747479761c9 req-d3eaeb11-7f71-4c35-8bcc-76533385355a service nova] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 688.453399] env[62692]: DEBUG nova.network.neutron [-] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.564096] env[62692]: DEBUG nova.network.neutron [req-491d5e80-a905-429d-bee8-6747479761c9 req-d3eaeb11-7f71-4c35-8bcc-76533385355a service nova] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.602333] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.080s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.602964] env[62692]: ERROR nova.compute.manager [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4122978e-21d6-422d-a286-589614e44e7c, please check neutron logs for more information. [ 688.602964] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Traceback (most recent call last): [ 688.602964] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 688.602964] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] self.driver.spawn(context, instance, image_meta, [ 688.602964] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 688.602964] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 688.602964] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 688.602964] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] vm_ref = self.build_virtual_machine(instance, [ 688.602964] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 688.602964] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] vif_infos = vmwarevif.get_vif_info(self._session, [ 688.602964] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 688.603294] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] for vif in network_info: [ 688.603294] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 688.603294] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] return self._sync_wrapper(fn, *args, **kwargs) [ 688.603294] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 688.603294] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] self.wait() [ 688.603294] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 688.603294] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] self[:] = self._gt.wait() [ 688.603294] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 688.603294] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] return self._exit_event.wait() [ 688.603294] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 688.603294] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] result = hub.switch() [ 688.603294] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 688.603294] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] return self.greenlet.switch() [ 688.603622] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 688.603622] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] result = function(*args, **kwargs) [ 688.603622] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 688.603622] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] return func(*args, **kwargs) [ 688.603622] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 688.603622] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] raise e [ 688.603622] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 688.603622] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] nwinfo = self.network_api.allocate_for_instance( [ 688.603622] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 688.603622] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] created_port_ids = self._update_ports_for_instance( [ 688.603622] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 688.603622] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] with excutils.save_and_reraise_exception(): [ 688.603622] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 688.603969] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] self.force_reraise() [ 688.603969] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 688.603969] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] raise self.value [ 688.603969] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 688.603969] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] updated_port = self._update_port( [ 688.603969] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 688.603969] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] _ensure_no_port_binding_failure(port) [ 688.603969] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 688.603969] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] raise exception.PortBindingFailed(port_id=port['id']) [ 688.603969] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] nova.exception.PortBindingFailed: Binding failed for port 4122978e-21d6-422d-a286-589614e44e7c, please check neutron logs for more information. [ 688.603969] env[62692]: ERROR nova.compute.manager [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] [ 688.604267] env[62692]: DEBUG nova.compute.utils [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Binding failed for port 4122978e-21d6-422d-a286-589614e44e7c, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 688.605846] env[62692]: DEBUG oslo_concurrency.lockutils [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.961s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.610926] env[62692]: DEBUG nova.compute.manager [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Build of instance e6a36771-c121-4ff5-8a78-c7b7868bf95b was re-scheduled: Binding failed for port 4122978e-21d6-422d-a286-589614e44e7c, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 688.611229] env[62692]: DEBUG nova.compute.manager [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 688.611510] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Acquiring lock "refresh_cache-e6a36771-c121-4ff5-8a78-c7b7868bf95b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.611610] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Acquired lock "refresh_cache-e6a36771-c121-4ff5-8a78-c7b7868bf95b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.611682] env[62692]: DEBUG nova.network.neutron [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 688.959891] env[62692]: INFO nova.compute.manager [-] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Took 1.05 seconds to deallocate network for instance. [ 688.961388] env[62692]: DEBUG nova.compute.claims [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 688.962328] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.067728] env[62692]: DEBUG oslo_concurrency.lockutils [req-491d5e80-a905-429d-bee8-6747479761c9 req-d3eaeb11-7f71-4c35-8bcc-76533385355a service nova] Releasing lock "refresh_cache-5d18922d-5937-47c8-a04a-aa757c153e8f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.067728] env[62692]: DEBUG nova.compute.manager [req-491d5e80-a905-429d-bee8-6747479761c9 req-d3eaeb11-7f71-4c35-8bcc-76533385355a service nova] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Received event network-vif-deleted-501c0ab4-2706-44d5-9057-d5ae4b34ae1f {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 689.137715] env[62692]: DEBUG nova.network.neutron [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 689.242425] env[62692]: DEBUG nova.network.neutron [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.426282] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Acquiring lock "1cdd1dbc-567d-46f6-9f05-dfbc57acbf99" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.426282] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Lock "1cdd1dbc-567d-46f6-9f05-dfbc57acbf99" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.687898] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3651c7eb-6e42-4b1e-a751-1fda38d6dde5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.697199] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4718b669-05a1-4f4e-bd5d-01b3ff79e0b1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.726980] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9b9c210-3243-4f20-90a0-6ee94ed7e865 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.734521] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e803d8b-c71b-4d7d-b24f-bdcf7e3929f2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.747879] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Releasing lock "refresh_cache-e6a36771-c121-4ff5-8a78-c7b7868bf95b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.748125] env[62692]: DEBUG nova.compute.manager [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 689.748299] env[62692]: DEBUG nova.compute.manager [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 689.748462] env[62692]: DEBUG nova.network.neutron [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 689.751833] env[62692]: DEBUG nova.compute.provider_tree [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 689.769893] env[62692]: DEBUG nova.network.neutron [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 690.254605] env[62692]: DEBUG nova.scheduler.client.report [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 690.273247] env[62692]: DEBUG nova.network.neutron [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.759420] env[62692]: DEBUG oslo_concurrency.lockutils [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.154s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.759987] env[62692]: ERROR nova.compute.manager [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e58be796-e8e0-4b09-a0bf-334afc678ee1, please check neutron logs for more information. [ 690.759987] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Traceback (most recent call last): [ 690.759987] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 690.759987] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] self.driver.spawn(context, instance, image_meta, [ 690.759987] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 690.759987] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] self._vmops.spawn(context, instance, image_meta, injected_files, [ 690.759987] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 690.759987] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] vm_ref = self.build_virtual_machine(instance, [ 690.759987] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 690.759987] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] vif_infos = vmwarevif.get_vif_info(self._session, [ 690.759987] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 690.760480] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] for vif in network_info: [ 690.760480] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 690.760480] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] return self._sync_wrapper(fn, *args, **kwargs) [ 690.760480] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 690.760480] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] self.wait() [ 690.760480] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 690.760480] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] self[:] = self._gt.wait() [ 690.760480] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 690.760480] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] return self._exit_event.wait() [ 690.760480] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 690.760480] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] result = hub.switch() [ 690.760480] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 690.760480] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] return self.greenlet.switch() [ 690.760824] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 690.760824] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] result = function(*args, **kwargs) [ 690.760824] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 690.760824] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] return func(*args, **kwargs) [ 690.760824] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 690.760824] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] raise e [ 690.760824] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 690.760824] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] nwinfo = self.network_api.allocate_for_instance( [ 690.760824] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 690.760824] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] created_port_ids = self._update_ports_for_instance( [ 690.760824] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 690.760824] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] with excutils.save_and_reraise_exception(): [ 690.760824] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 690.761116] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] self.force_reraise() [ 690.761116] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 690.761116] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] raise self.value [ 690.761116] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 690.761116] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] updated_port = self._update_port( [ 690.761116] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 690.761116] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] _ensure_no_port_binding_failure(port) [ 690.761116] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 690.761116] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] raise exception.PortBindingFailed(port_id=port['id']) [ 690.761116] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] nova.exception.PortBindingFailed: Binding failed for port e58be796-e8e0-4b09-a0bf-334afc678ee1, please check neutron logs for more information. [ 690.761116] env[62692]: ERROR nova.compute.manager [instance: 891170e7-35db-4faa-a1be-fcf376faee93] [ 690.761384] env[62692]: DEBUG nova.compute.utils [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Binding failed for port e58be796-e8e0-4b09-a0bf-334afc678ee1, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 690.762538] env[62692]: DEBUG nova.compute.manager [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Build of instance 891170e7-35db-4faa-a1be-fcf376faee93 was re-scheduled: Binding failed for port e58be796-e8e0-4b09-a0bf-334afc678ee1, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 690.763857] env[62692]: DEBUG nova.compute.manager [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 690.763857] env[62692]: DEBUG oslo_concurrency.lockutils [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Acquiring lock "refresh_cache-891170e7-35db-4faa-a1be-fcf376faee93" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.763857] env[62692]: DEBUG oslo_concurrency.lockutils [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Acquired lock "refresh_cache-891170e7-35db-4faa-a1be-fcf376faee93" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.763857] env[62692]: DEBUG nova.network.neutron [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 690.764991] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.600s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.766565] env[62692]: INFO nova.compute.claims [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 690.777451] env[62692]: INFO nova.compute.manager [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] [instance: e6a36771-c121-4ff5-8a78-c7b7868bf95b] Took 1.03 seconds to deallocate network for instance. [ 691.292893] env[62692]: DEBUG nova.network.neutron [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 691.407008] env[62692]: DEBUG nova.network.neutron [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.812547] env[62692]: INFO nova.scheduler.client.report [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Deleted allocations for instance e6a36771-c121-4ff5-8a78-c7b7868bf95b [ 691.911118] env[62692]: DEBUG oslo_concurrency.lockutils [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Releasing lock "refresh_cache-891170e7-35db-4faa-a1be-fcf376faee93" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.911307] env[62692]: DEBUG nova.compute.manager [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 691.911433] env[62692]: DEBUG nova.compute.manager [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 691.911600] env[62692]: DEBUG nova.network.neutron [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 691.943574] env[62692]: DEBUG nova.network.neutron [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 692.294493] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9913ec8-abe7-4355-98f8-0fe389630b98 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.302321] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a6e004e-b19c-4720-b9bf-d6d660d2811a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.333046] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b6aae002-5496-40bb-8e5d-328a6e78a0d6 tempest-InstanceActionsV221TestJSON-607505560 tempest-InstanceActionsV221TestJSON-607505560-project-member] Lock "e6a36771-c121-4ff5-8a78-c7b7868bf95b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.320s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.334928] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfb1ba28-78a1-4e47-bc66-b2844f89bc92 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.343914] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dcd8c5f-1477-45cf-8056-e94b43cd7339 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.361021] env[62692]: DEBUG nova.compute.provider_tree [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 692.445898] env[62692]: DEBUG nova.network.neutron [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.839909] env[62692]: DEBUG nova.compute.manager [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 692.863860] env[62692]: DEBUG nova.scheduler.client.report [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 692.948847] env[62692]: INFO nova.compute.manager [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] [instance: 891170e7-35db-4faa-a1be-fcf376faee93] Took 1.04 seconds to deallocate network for instance. [ 693.365979] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.372279] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.605s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.372279] env[62692]: DEBUG nova.compute.manager [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 693.373709] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.243s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.883196] env[62692]: DEBUG nova.compute.utils [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 693.889654] env[62692]: DEBUG nova.compute.manager [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 693.894066] env[62692]: DEBUG nova.network.neutron [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 693.960098] env[62692]: DEBUG nova.policy [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f2e2ee26163841ddb15bb67520e98394', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e13f8888d5e84eed8247101558e29a56', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 693.991414] env[62692]: INFO nova.scheduler.client.report [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Deleted allocations for instance 891170e7-35db-4faa-a1be-fcf376faee93 [ 694.383856] env[62692]: DEBUG nova.network.neutron [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Successfully created port: 451c23dc-3dfb-4dc8-9e9b-eb70351b81bb {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 694.388873] env[62692]: DEBUG nova.compute.manager [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 694.446095] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39bbbcba-5ded-473a-bae7-59d16de32bb2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.453266] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-135a6bdc-ef8c-4d0f-943e-fcf71c058743 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.485504] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f9d9125-6fa9-47b4-b434-e84d0ce9d91e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.493500] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-728cb9e2-0422-4d61-a425-42286b89ce7d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.508016] env[62692]: DEBUG oslo_concurrency.lockutils [None req-40a92707-303c-434b-a7bf-37feb287d510 tempest-ServersWithSpecificFlavorTestJSON-1051974771 tempest-ServersWithSpecificFlavorTestJSON-1051974771-project-member] Lock "891170e7-35db-4faa-a1be-fcf376faee93" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.164s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.508016] env[62692]: DEBUG nova.compute.provider_tree [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 695.012520] env[62692]: DEBUG nova.scheduler.client.report [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 695.015721] env[62692]: DEBUG nova.compute.manager [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 695.403837] env[62692]: DEBUG nova.compute.manager [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 695.434173] env[62692]: DEBUG nova.virt.hardware [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 695.434173] env[62692]: DEBUG nova.virt.hardware [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 695.434173] env[62692]: DEBUG nova.virt.hardware [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 695.434173] env[62692]: DEBUG nova.virt.hardware [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 695.434356] env[62692]: DEBUG nova.virt.hardware [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 695.434356] env[62692]: DEBUG nova.virt.hardware [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 695.434356] env[62692]: DEBUG nova.virt.hardware [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 695.434356] env[62692]: DEBUG nova.virt.hardware [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 695.434356] env[62692]: DEBUG nova.virt.hardware [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 695.434517] env[62692]: DEBUG nova.virt.hardware [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 695.434517] env[62692]: DEBUG nova.virt.hardware [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 695.434517] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57ce4b64-0884-4d2b-9825-72f620909348 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.442948] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d9c877d-b1a8-4d9b-ab07-0e9a1b5d3164 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.521590] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.148s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.522276] env[62692]: ERROR nova.compute.manager [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 44e32753-d6cd-4e32-8b26-9476d40e482e, please check neutron logs for more information. [ 695.522276] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Traceback (most recent call last): [ 695.522276] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 695.522276] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] self.driver.spawn(context, instance, image_meta, [ 695.522276] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 695.522276] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 695.522276] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 695.522276] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] vm_ref = self.build_virtual_machine(instance, [ 695.522276] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 695.522276] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] vif_infos = vmwarevif.get_vif_info(self._session, [ 695.522276] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 695.522656] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] for vif in network_info: [ 695.522656] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 695.522656] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] return self._sync_wrapper(fn, *args, **kwargs) [ 695.522656] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 695.522656] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] self.wait() [ 695.522656] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 695.522656] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] self[:] = self._gt.wait() [ 695.522656] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 695.522656] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] return self._exit_event.wait() [ 695.522656] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 695.522656] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] result = hub.switch() [ 695.522656] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 695.522656] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] return self.greenlet.switch() [ 695.523031] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 695.523031] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] result = function(*args, **kwargs) [ 695.523031] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 695.523031] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] return func(*args, **kwargs) [ 695.523031] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 695.523031] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] raise e [ 695.523031] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 695.523031] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] nwinfo = self.network_api.allocate_for_instance( [ 695.523031] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 695.523031] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] created_port_ids = self._update_ports_for_instance( [ 695.523031] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 695.523031] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] with excutils.save_and_reraise_exception(): [ 695.523031] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.523395] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] self.force_reraise() [ 695.523395] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.523395] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] raise self.value [ 695.523395] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 695.523395] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] updated_port = self._update_port( [ 695.523395] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.523395] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] _ensure_no_port_binding_failure(port) [ 695.523395] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.523395] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] raise exception.PortBindingFailed(port_id=port['id']) [ 695.523395] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] nova.exception.PortBindingFailed: Binding failed for port 44e32753-d6cd-4e32-8b26-9476d40e482e, please check neutron logs for more information. [ 695.523395] env[62692]: ERROR nova.compute.manager [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] [ 695.523716] env[62692]: DEBUG nova.compute.utils [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Binding failed for port 44e32753-d6cd-4e32-8b26-9476d40e482e, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 695.530713] env[62692]: DEBUG nova.compute.manager [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Build of instance 59882012-b01c-45ec-bd53-6e17e356ce5f was re-scheduled: Binding failed for port 44e32753-d6cd-4e32-8b26-9476d40e482e, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 695.531194] env[62692]: DEBUG nova.compute.manager [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 695.531425] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Acquiring lock "refresh_cache-59882012-b01c-45ec-bd53-6e17e356ce5f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.531632] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Acquired lock "refresh_cache-59882012-b01c-45ec-bd53-6e17e356ce5f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.531722] env[62692]: DEBUG nova.network.neutron [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 695.532907] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.228s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.553363] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.834972] env[62692]: ERROR nova.compute.manager [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 451c23dc-3dfb-4dc8-9e9b-eb70351b81bb, please check neutron logs for more information. [ 695.834972] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 695.834972] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 695.834972] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 695.834972] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 695.834972] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 695.834972] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 695.834972] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 695.834972] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.834972] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 695.834972] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.834972] env[62692]: ERROR nova.compute.manager raise self.value [ 695.834972] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 695.834972] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 695.834972] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.834972] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 695.835389] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.835389] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 695.835389] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 451c23dc-3dfb-4dc8-9e9b-eb70351b81bb, please check neutron logs for more information. [ 695.835389] env[62692]: ERROR nova.compute.manager [ 695.835389] env[62692]: Traceback (most recent call last): [ 695.835389] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 695.835389] env[62692]: listener.cb(fileno) [ 695.835389] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 695.835389] env[62692]: result = function(*args, **kwargs) [ 695.835389] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 695.835389] env[62692]: return func(*args, **kwargs) [ 695.835389] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 695.835389] env[62692]: raise e [ 695.835389] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 695.835389] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 695.835389] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 695.835389] env[62692]: created_port_ids = self._update_ports_for_instance( [ 695.835389] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 695.835389] env[62692]: with excutils.save_and_reraise_exception(): [ 695.835389] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.835389] env[62692]: self.force_reraise() [ 695.835389] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.835389] env[62692]: raise self.value [ 695.835389] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 695.835389] env[62692]: updated_port = self._update_port( [ 695.835389] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.835389] env[62692]: _ensure_no_port_binding_failure(port) [ 695.835389] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.835389] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 695.836051] env[62692]: nova.exception.PortBindingFailed: Binding failed for port 451c23dc-3dfb-4dc8-9e9b-eb70351b81bb, please check neutron logs for more information. [ 695.836051] env[62692]: Removing descriptor: 19 [ 695.836051] env[62692]: ERROR nova.compute.manager [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 451c23dc-3dfb-4dc8-9e9b-eb70351b81bb, please check neutron logs for more information. [ 695.836051] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Traceback (most recent call last): [ 695.836051] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 695.836051] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] yield resources [ 695.836051] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 695.836051] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] self.driver.spawn(context, instance, image_meta, [ 695.836051] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 695.836051] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 695.836051] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 695.836051] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] vm_ref = self.build_virtual_machine(instance, [ 695.836342] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 695.836342] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] vif_infos = vmwarevif.get_vif_info(self._session, [ 695.836342] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 695.836342] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] for vif in network_info: [ 695.836342] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 695.836342] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] return self._sync_wrapper(fn, *args, **kwargs) [ 695.836342] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 695.836342] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] self.wait() [ 695.836342] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 695.836342] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] self[:] = self._gt.wait() [ 695.836342] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 695.836342] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] return self._exit_event.wait() [ 695.836342] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 695.836628] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] result = hub.switch() [ 695.836628] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 695.836628] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] return self.greenlet.switch() [ 695.836628] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 695.836628] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] result = function(*args, **kwargs) [ 695.836628] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 695.836628] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] return func(*args, **kwargs) [ 695.836628] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 695.836628] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] raise e [ 695.836628] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 695.836628] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] nwinfo = self.network_api.allocate_for_instance( [ 695.836628] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 695.836628] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] created_port_ids = self._update_ports_for_instance( [ 695.836932] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 695.836932] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] with excutils.save_and_reraise_exception(): [ 695.836932] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.836932] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] self.force_reraise() [ 695.836932] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.836932] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] raise self.value [ 695.836932] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 695.836932] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] updated_port = self._update_port( [ 695.836932] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.836932] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] _ensure_no_port_binding_failure(port) [ 695.836932] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.836932] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] raise exception.PortBindingFailed(port_id=port['id']) [ 695.837203] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] nova.exception.PortBindingFailed: Binding failed for port 451c23dc-3dfb-4dc8-9e9b-eb70351b81bb, please check neutron logs for more information. [ 695.837203] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] [ 695.837203] env[62692]: INFO nova.compute.manager [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Terminating instance [ 695.838236] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "refresh_cache-b76cce5f-b4ab-424f-b604-684f5c22154a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.838389] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquired lock "refresh_cache-b76cce5f-b4ab-424f-b604-684f5c22154a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.838554] env[62692]: DEBUG nova.network.neutron [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 695.916615] env[62692]: DEBUG nova.compute.manager [req-f6a3a879-adba-415a-a842-1d2c48fa5487 req-1eacf799-1734-4d1d-8478-d35e6ca141db service nova] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Received event network-changed-451c23dc-3dfb-4dc8-9e9b-eb70351b81bb {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 695.916615] env[62692]: DEBUG nova.compute.manager [req-f6a3a879-adba-415a-a842-1d2c48fa5487 req-1eacf799-1734-4d1d-8478-d35e6ca141db service nova] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Refreshing instance network info cache due to event network-changed-451c23dc-3dfb-4dc8-9e9b-eb70351b81bb. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 695.916815] env[62692]: DEBUG oslo_concurrency.lockutils [req-f6a3a879-adba-415a-a842-1d2c48fa5487 req-1eacf799-1734-4d1d-8478-d35e6ca141db service nova] Acquiring lock "refresh_cache-b76cce5f-b4ab-424f-b604-684f5c22154a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.060746] env[62692]: DEBUG nova.network.neutron [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 696.150710] env[62692]: DEBUG nova.network.neutron [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.368390] env[62692]: DEBUG nova.network.neutron [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 696.509645] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4a1583c-0b41-4e10-ac41-00dd12b654d5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.517798] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9665ba75-d1c1-4b5c-aaa3-187cb0b24836 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.523540] env[62692]: DEBUG nova.network.neutron [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.560313] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5367778-68c1-424d-8720-41660f608210 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.570094] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eeb2d52-99bb-40a4-b8f8-9ba7d6a603ba {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.587175] env[62692]: DEBUG nova.compute.provider_tree [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 696.653222] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Releasing lock "refresh_cache-59882012-b01c-45ec-bd53-6e17e356ce5f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.653712] env[62692]: DEBUG nova.compute.manager [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 696.654013] env[62692]: DEBUG nova.compute.manager [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 696.654204] env[62692]: DEBUG nova.network.neutron [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 696.673357] env[62692]: DEBUG nova.network.neutron [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 697.026862] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Releasing lock "refresh_cache-b76cce5f-b4ab-424f-b604-684f5c22154a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.027330] env[62692]: DEBUG nova.compute.manager [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 697.027523] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 697.027977] env[62692]: DEBUG oslo_concurrency.lockutils [req-f6a3a879-adba-415a-a842-1d2c48fa5487 req-1eacf799-1734-4d1d-8478-d35e6ca141db service nova] Acquired lock "refresh_cache-b76cce5f-b4ab-424f-b604-684f5c22154a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.028101] env[62692]: DEBUG nova.network.neutron [req-f6a3a879-adba-415a-a842-1d2c48fa5487 req-1eacf799-1734-4d1d-8478-d35e6ca141db service nova] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Refreshing network info cache for port 451c23dc-3dfb-4dc8-9e9b-eb70351b81bb {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 697.029545] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-acbc04a5-02bc-4097-bd01-c987057cd83c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.039573] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4354feb6-2b32-467b-b555-1fa0c0e0a912 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.063914] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b76cce5f-b4ab-424f-b604-684f5c22154a could not be found. [ 697.064187] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 697.064372] env[62692]: INFO nova.compute.manager [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 697.064614] env[62692]: DEBUG oslo.service.loopingcall [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 697.064825] env[62692]: DEBUG nova.compute.manager [-] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 697.064937] env[62692]: DEBUG nova.network.neutron [-] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 697.087619] env[62692]: DEBUG nova.network.neutron [-] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 697.092059] env[62692]: DEBUG nova.scheduler.client.report [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 697.175866] env[62692]: DEBUG nova.network.neutron [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.554306] env[62692]: DEBUG nova.network.neutron [req-f6a3a879-adba-415a-a842-1d2c48fa5487 req-1eacf799-1734-4d1d-8478-d35e6ca141db service nova] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 697.598114] env[62692]: DEBUG nova.network.neutron [-] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.598114] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.064s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.598114] env[62692]: ERROR nova.compute.manager [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c049b809-696e-4042-ada6-74917ad431b6, please check neutron logs for more information. [ 697.598114] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Traceback (most recent call last): [ 697.598114] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 697.598114] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] self.driver.spawn(context, instance, image_meta, [ 697.598114] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 697.598114] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 697.598114] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 697.598447] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] vm_ref = self.build_virtual_machine(instance, [ 697.598447] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 697.598447] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] vif_infos = vmwarevif.get_vif_info(self._session, [ 697.598447] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 697.598447] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] for vif in network_info: [ 697.598447] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 697.598447] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] return self._sync_wrapper(fn, *args, **kwargs) [ 697.598447] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 697.598447] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] self.wait() [ 697.598447] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 697.598447] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] self[:] = self._gt.wait() [ 697.598447] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 697.598447] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] return self._exit_event.wait() [ 697.598812] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 697.598812] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] result = hub.switch() [ 697.598812] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 697.598812] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] return self.greenlet.switch() [ 697.598812] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 697.598812] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] result = function(*args, **kwargs) [ 697.598812] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 697.598812] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] return func(*args, **kwargs) [ 697.598812] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 697.598812] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] raise e [ 697.598812] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 697.598812] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] nwinfo = self.network_api.allocate_for_instance( [ 697.598812] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 697.599166] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] created_port_ids = self._update_ports_for_instance( [ 697.599166] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 697.599166] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] with excutils.save_and_reraise_exception(): [ 697.599166] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.599166] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] self.force_reraise() [ 697.599166] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.599166] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] raise self.value [ 697.599166] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 697.599166] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] updated_port = self._update_port( [ 697.599166] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.599166] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] _ensure_no_port_binding_failure(port) [ 697.599166] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.599166] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] raise exception.PortBindingFailed(port_id=port['id']) [ 697.599499] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] nova.exception.PortBindingFailed: Binding failed for port c049b809-696e-4042-ada6-74917ad431b6, please check neutron logs for more information. [ 697.599499] env[62692]: ERROR nova.compute.manager [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] [ 697.599499] env[62692]: DEBUG nova.compute.utils [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Binding failed for port c049b809-696e-4042-ada6-74917ad431b6, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 697.602211] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.800s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.608161] env[62692]: DEBUG nova.compute.manager [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Build of instance b014c45b-4d27-4ade-b823-0dd64dfea2e2 was re-scheduled: Binding failed for port c049b809-696e-4042-ada6-74917ad431b6, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 697.608161] env[62692]: DEBUG nova.compute.manager [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 697.608161] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Acquiring lock "refresh_cache-b014c45b-4d27-4ade-b823-0dd64dfea2e2" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 697.608161] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Acquired lock "refresh_cache-b014c45b-4d27-4ade-b823-0dd64dfea2e2" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.608161] env[62692]: DEBUG nova.network.neutron [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 697.664880] env[62692]: DEBUG nova.network.neutron [req-f6a3a879-adba-415a-a842-1d2c48fa5487 req-1eacf799-1734-4d1d-8478-d35e6ca141db service nova] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.678412] env[62692]: INFO nova.compute.manager [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] [instance: 59882012-b01c-45ec-bd53-6e17e356ce5f] Took 1.02 seconds to deallocate network for instance. [ 698.103891] env[62692]: INFO nova.compute.manager [-] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Took 1.04 seconds to deallocate network for instance. [ 698.104907] env[62692]: DEBUG nova.compute.claims [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 698.104907] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.108055] env[62692]: DEBUG nova.compute.manager [req-b3577019-4fb5-48bb-83e7-85530bc28d9c req-e6b0ec20-6cc3-40bb-8918-6486db6b1b02 service nova] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Received event network-vif-deleted-451c23dc-3dfb-4dc8-9e9b-eb70351b81bb {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 698.147920] env[62692]: DEBUG nova.network.neutron [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 698.171495] env[62692]: DEBUG oslo_concurrency.lockutils [req-f6a3a879-adba-415a-a842-1d2c48fa5487 req-1eacf799-1734-4d1d-8478-d35e6ca141db service nova] Releasing lock "refresh_cache-b76cce5f-b4ab-424f-b604-684f5c22154a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.254654] env[62692]: DEBUG nova.network.neutron [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.664706] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-932af4e5-977b-4745-bc51-538d4c8ad82a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.676049] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd072587-9eb4-47a6-8788-85347508dd4d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.708382] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1e02b6c-93d9-4213-a028-6b16df9f12b3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.716448] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1622d58e-ffc6-4780-92f8-f55eccca335b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.724045] env[62692]: INFO nova.scheduler.client.report [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Deleted allocations for instance 59882012-b01c-45ec-bd53-6e17e356ce5f [ 698.737788] env[62692]: DEBUG nova.compute.provider_tree [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 698.758490] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Releasing lock "refresh_cache-b014c45b-4d27-4ade-b823-0dd64dfea2e2" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.758610] env[62692]: DEBUG nova.compute.manager [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 698.758882] env[62692]: DEBUG nova.compute.manager [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 698.761806] env[62692]: DEBUG nova.network.neutron [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 698.784821] env[62692]: DEBUG nova.network.neutron [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 699.241043] env[62692]: DEBUG nova.scheduler.client.report [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 699.244335] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9600f6e8-bf77-4369-8875-8f57bae6e9e8 tempest-InstanceActionsTestJSON-2090305130 tempest-InstanceActionsTestJSON-2090305130-project-member] Lock "59882012-b01c-45ec-bd53-6e17e356ce5f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.651s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.288337] env[62692]: DEBUG nova.network.neutron [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.747248] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.146s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.747984] env[62692]: ERROR nova.compute.manager [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6c9a2c2d-decd-4dc2-b0af-1c77575dd5cd, please check neutron logs for more information. [ 699.747984] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Traceback (most recent call last): [ 699.747984] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 699.747984] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] self.driver.spawn(context, instance, image_meta, [ 699.747984] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 699.747984] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 699.747984] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 699.747984] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] vm_ref = self.build_virtual_machine(instance, [ 699.747984] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 699.747984] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] vif_infos = vmwarevif.get_vif_info(self._session, [ 699.747984] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 699.748285] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] for vif in network_info: [ 699.748285] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 699.748285] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] return self._sync_wrapper(fn, *args, **kwargs) [ 699.748285] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 699.748285] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] self.wait() [ 699.748285] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 699.748285] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] self[:] = self._gt.wait() [ 699.748285] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 699.748285] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] return self._exit_event.wait() [ 699.748285] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 699.748285] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] result = hub.switch() [ 699.748285] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 699.748285] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] return self.greenlet.switch() [ 699.748588] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.748588] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] result = function(*args, **kwargs) [ 699.748588] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 699.748588] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] return func(*args, **kwargs) [ 699.748588] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.748588] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] raise e [ 699.748588] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.748588] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] nwinfo = self.network_api.allocate_for_instance( [ 699.748588] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 699.748588] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] created_port_ids = self._update_ports_for_instance( [ 699.748588] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 699.748588] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] with excutils.save_and_reraise_exception(): [ 699.748588] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.748904] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] self.force_reraise() [ 699.748904] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.748904] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] raise self.value [ 699.748904] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 699.748904] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] updated_port = self._update_port( [ 699.748904] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.748904] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] _ensure_no_port_binding_failure(port) [ 699.748904] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.748904] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] raise exception.PortBindingFailed(port_id=port['id']) [ 699.748904] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] nova.exception.PortBindingFailed: Binding failed for port 6c9a2c2d-decd-4dc2-b0af-1c77575dd5cd, please check neutron logs for more information. [ 699.748904] env[62692]: ERROR nova.compute.manager [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] [ 699.750387] env[62692]: DEBUG nova.compute.utils [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Binding failed for port 6c9a2c2d-decd-4dc2-b0af-1c77575dd5cd, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 699.751154] env[62692]: DEBUG nova.compute.manager [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 699.754952] env[62692]: DEBUG nova.compute.manager [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Build of instance 6883c19f-1e3a-4eac-be74-8e725faee77a was re-scheduled: Binding failed for port 6c9a2c2d-decd-4dc2-b0af-1c77575dd5cd, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 699.756513] env[62692]: DEBUG nova.compute.manager [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 699.756513] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Acquiring lock "refresh_cache-6883c19f-1e3a-4eac-be74-8e725faee77a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.756513] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Acquired lock "refresh_cache-6883c19f-1e3a-4eac-be74-8e725faee77a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.756513] env[62692]: DEBUG nova.network.neutron [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 699.757046] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.409s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.791837] env[62692]: INFO nova.compute.manager [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] [instance: b014c45b-4d27-4ade-b823-0dd64dfea2e2] Took 1.03 seconds to deallocate network for instance. [ 700.296563] env[62692]: DEBUG nova.network.neutron [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 700.309014] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 700.520803] env[62692]: DEBUG nova.network.neutron [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.801092] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d1e631f-7cd0-4e40-95af-61f5b345b864 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.809637] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a7aa0a8-76cb-42fb-a013-f9ad6e7bbb0a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.852343] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bb1ff5f-1bc5-431b-884c-7ba0af21a173 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.855836] env[62692]: INFO nova.scheduler.client.report [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Deleted allocations for instance b014c45b-4d27-4ade-b823-0dd64dfea2e2 [ 700.868238] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3c087a6-5d4a-45d5-8673-4473796b2de8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.882856] env[62692]: DEBUG nova.compute.provider_tree [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 701.024883] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Releasing lock "refresh_cache-6883c19f-1e3a-4eac-be74-8e725faee77a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.025184] env[62692]: DEBUG nova.compute.manager [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 701.025378] env[62692]: DEBUG nova.compute.manager [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 701.025549] env[62692]: DEBUG nova.network.neutron [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 701.052811] env[62692]: DEBUG nova.network.neutron [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 701.374875] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b718655-bc80-4d6e-b779-72be969b9c93 tempest-ServersTestJSON-626146439 tempest-ServersTestJSON-626146439-project-member] Lock "b014c45b-4d27-4ade-b823-0dd64dfea2e2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.064s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.387748] env[62692]: DEBUG nova.scheduler.client.report [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 701.555876] env[62692]: DEBUG nova.network.neutron [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.876407] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 701.896658] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.139s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.898595] env[62692]: ERROR nova.compute.manager [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cd7a8a64-e3fd-4852-b34b-f93d8452dd9a, please check neutron logs for more information. [ 701.898595] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] Traceback (most recent call last): [ 701.898595] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 701.898595] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] self.driver.spawn(context, instance, image_meta, [ 701.898595] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 701.898595] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 701.898595] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 701.898595] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] vm_ref = self.build_virtual_machine(instance, [ 701.898595] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 701.898595] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] vif_infos = vmwarevif.get_vif_info(self._session, [ 701.898595] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 701.901445] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] for vif in network_info: [ 701.901445] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 701.901445] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] return self._sync_wrapper(fn, *args, **kwargs) [ 701.901445] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 701.901445] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] self.wait() [ 701.901445] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 701.901445] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] self[:] = self._gt.wait() [ 701.901445] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 701.901445] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] return self._exit_event.wait() [ 701.901445] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 701.901445] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] result = hub.switch() [ 701.901445] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 701.901445] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] return self.greenlet.switch() [ 701.902085] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.902085] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] result = function(*args, **kwargs) [ 701.902085] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 701.902085] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] return func(*args, **kwargs) [ 701.902085] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 701.902085] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] raise e [ 701.902085] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.902085] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] nwinfo = self.network_api.allocate_for_instance( [ 701.902085] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 701.902085] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] created_port_ids = self._update_ports_for_instance( [ 701.902085] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 701.902085] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] with excutils.save_and_reraise_exception(): [ 701.902085] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.902448] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] self.force_reraise() [ 701.902448] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.902448] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] raise self.value [ 701.902448] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 701.902448] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] updated_port = self._update_port( [ 701.902448] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.902448] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] _ensure_no_port_binding_failure(port) [ 701.902448] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.902448] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] raise exception.PortBindingFailed(port_id=port['id']) [ 701.902448] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] nova.exception.PortBindingFailed: Binding failed for port cd7a8a64-e3fd-4852-b34b-f93d8452dd9a, please check neutron logs for more information. [ 701.902448] env[62692]: ERROR nova.compute.manager [instance: 844b3803-8a01-484e-8908-dbae27188acb] [ 701.902717] env[62692]: DEBUG nova.compute.utils [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Binding failed for port cd7a8a64-e3fd-4852-b34b-f93d8452dd9a, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 701.902717] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.904488] env[62692]: INFO nova.compute.claims [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 701.907965] env[62692]: DEBUG nova.compute.manager [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Build of instance 844b3803-8a01-484e-8908-dbae27188acb was re-scheduled: Binding failed for port cd7a8a64-e3fd-4852-b34b-f93d8452dd9a, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 701.908481] env[62692]: DEBUG nova.compute.manager [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 701.908708] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Acquiring lock "refresh_cache-844b3803-8a01-484e-8908-dbae27188acb" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.908865] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Acquired lock "refresh_cache-844b3803-8a01-484e-8908-dbae27188acb" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.909045] env[62692]: DEBUG nova.network.neutron [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 702.061276] env[62692]: INFO nova.compute.manager [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] [instance: 6883c19f-1e3a-4eac-be74-8e725faee77a] Took 1.03 seconds to deallocate network for instance. [ 702.402480] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.457986] env[62692]: DEBUG nova.network.neutron [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 702.583797] env[62692]: DEBUG nova.network.neutron [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.089435] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Releasing lock "refresh_cache-844b3803-8a01-484e-8908-dbae27188acb" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.089700] env[62692]: DEBUG nova.compute.manager [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 703.089896] env[62692]: DEBUG nova.compute.manager [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 703.090118] env[62692]: DEBUG nova.network.neutron [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 703.099944] env[62692]: INFO nova.scheduler.client.report [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Deleted allocations for instance 6883c19f-1e3a-4eac-be74-8e725faee77a [ 703.138984] env[62692]: DEBUG nova.network.neutron [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 703.433570] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eeb84cf-dd2c-4c43-b166-3715a44d3c4e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.442148] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52264d8a-290d-4e6a-9ddb-49ecc7115366 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.477589] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8874652f-1977-4e92-82ed-78b67b9a6c00 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.488407] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba4068b-aa86-4150-b75c-685c51cea806 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.507463] env[62692]: DEBUG nova.compute.provider_tree [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 703.613707] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6bdc8d6f-2b65-4d8b-868d-987dfceb8d51 tempest-ImagesOneServerNegativeTestJSON-1521504718 tempest-ImagesOneServerNegativeTestJSON-1521504718-project-member] Lock "6883c19f-1e3a-4eac-be74-8e725faee77a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.662s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.645667] env[62692]: DEBUG nova.network.neutron [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.014762] env[62692]: DEBUG nova.scheduler.client.report [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 704.118110] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 704.151050] env[62692]: INFO nova.compute.manager [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 844b3803-8a01-484e-8908-dbae27188acb] Took 1.06 seconds to deallocate network for instance. [ 704.520088] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.618s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.521142] env[62692]: DEBUG nova.compute.manager [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 704.524320] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.362s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.655423] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.033266] env[62692]: DEBUG nova.compute.utils [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 705.037131] env[62692]: DEBUG nova.compute.manager [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 705.037317] env[62692]: DEBUG nova.network.neutron [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 705.193609] env[62692]: DEBUG nova.policy [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de581ad331f8489285974b3fb3e38da8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1fc24cdd7a8c4e3481008f773242ecca', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 705.198275] env[62692]: INFO nova.scheduler.client.report [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Deleted allocations for instance 844b3803-8a01-484e-8908-dbae27188acb [ 705.538539] env[62692]: DEBUG nova.compute.manager [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 705.609127] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1ca3283-cfaa-4c34-96b9-eb522c4aae0a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.620837] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6f9c866-76cc-495a-81d2-b0bd62e8c956 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.657543] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfffff04-e0e0-4f64-9c25-256a6e7c02a7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.669634] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a646e2ae-9ec9-4893-9903-2fd6ed894052 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.686578] env[62692]: DEBUG nova.compute.provider_tree [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 705.716087] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1c10cef9-1055-41d4-bf3e-ace2f0ff9ad4 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Lock "844b3803-8a01-484e-8908-dbae27188acb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.904s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.191975] env[62692]: DEBUG nova.scheduler.client.report [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 706.218424] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 706.387994] env[62692]: DEBUG nova.network.neutron [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Successfully created port: eed49ef6-5628-4bd8-a95e-05d1f3c803cc {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 706.561891] env[62692]: DEBUG nova.compute.manager [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 706.604805] env[62692]: DEBUG nova.virt.hardware [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 706.604805] env[62692]: DEBUG nova.virt.hardware [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 706.605067] env[62692]: DEBUG nova.virt.hardware [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 706.605275] env[62692]: DEBUG nova.virt.hardware [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 706.605415] env[62692]: DEBUG nova.virt.hardware [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 706.605556] env[62692]: DEBUG nova.virt.hardware [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 706.606074] env[62692]: DEBUG nova.virt.hardware [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 706.606074] env[62692]: DEBUG nova.virt.hardware [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 706.606074] env[62692]: DEBUG nova.virt.hardware [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 706.606240] env[62692]: DEBUG nova.virt.hardware [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 706.606520] env[62692]: DEBUG nova.virt.hardware [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 706.607321] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bcc5154-1cf9-4c4b-a688-3a6701ada768 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.617640] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f7b35ce-b0b3-4cce-87ca-c023a7ee4655 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.702743] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.178s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.703435] env[62692]: ERROR nova.compute.manager [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 32670d72-1490-4454-aa34-853d890b202b, please check neutron logs for more information. [ 706.703435] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Traceback (most recent call last): [ 706.703435] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 706.703435] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] self.driver.spawn(context, instance, image_meta, [ 706.703435] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 706.703435] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] self._vmops.spawn(context, instance, image_meta, injected_files, [ 706.703435] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 706.703435] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] vm_ref = self.build_virtual_machine(instance, [ 706.703435] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 706.703435] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] vif_infos = vmwarevif.get_vif_info(self._session, [ 706.703435] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 706.703763] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] for vif in network_info: [ 706.703763] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 706.703763] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] return self._sync_wrapper(fn, *args, **kwargs) [ 706.703763] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 706.703763] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] self.wait() [ 706.703763] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 706.703763] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] self[:] = self._gt.wait() [ 706.703763] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 706.703763] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] return self._exit_event.wait() [ 706.703763] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 706.703763] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] current.throw(*self._exc) [ 706.703763] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 706.703763] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] result = function(*args, **kwargs) [ 706.704119] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 706.704119] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] return func(*args, **kwargs) [ 706.704119] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 706.704119] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] raise e [ 706.704119] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.704119] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] nwinfo = self.network_api.allocate_for_instance( [ 706.704119] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 706.704119] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] created_port_ids = self._update_ports_for_instance( [ 706.704119] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 706.704119] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] with excutils.save_and_reraise_exception(): [ 706.704119] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.704119] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] self.force_reraise() [ 706.704119] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.704559] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] raise self.value [ 706.704559] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 706.704559] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] updated_port = self._update_port( [ 706.704559] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.704559] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] _ensure_no_port_binding_failure(port) [ 706.704559] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.704559] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] raise exception.PortBindingFailed(port_id=port['id']) [ 706.704559] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] nova.exception.PortBindingFailed: Binding failed for port 32670d72-1490-4454-aa34-853d890b202b, please check neutron logs for more information. [ 706.704559] env[62692]: ERROR nova.compute.manager [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] [ 706.704559] env[62692]: DEBUG nova.compute.utils [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Binding failed for port 32670d72-1490-4454-aa34-853d890b202b, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 706.705917] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.744s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.708658] env[62692]: DEBUG nova.compute.manager [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Build of instance 0aac5bf9-2af4-4745-a961-46986fb57779 was re-scheduled: Binding failed for port 32670d72-1490-4454-aa34-853d890b202b, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 706.709284] env[62692]: DEBUG nova.compute.manager [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 706.709534] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Acquiring lock "refresh_cache-0aac5bf9-2af4-4745-a961-46986fb57779" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.709685] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Acquired lock "refresh_cache-0aac5bf9-2af4-4745-a961-46986fb57779" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.709845] env[62692]: DEBUG nova.network.neutron [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 706.753177] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.259312] env[62692]: DEBUG nova.network.neutron [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 707.476444] env[62692]: DEBUG oslo_concurrency.lockutils [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Acquiring lock "7a3f9c7d-03cf-4177-88db-acf33cc03feb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.476821] env[62692]: DEBUG oslo_concurrency.lockutils [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Lock "7a3f9c7d-03cf-4177-88db-acf33cc03feb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.578249] env[62692]: DEBUG nova.network.neutron [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.755985] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c42200d6-959d-4011-a474-2177d4b63fd9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.763756] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40aac217-55be-4468-8fae-593e9228da13 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.797380] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78c3dc79-b972-4f1b-bdf9-ee76c7f84b01 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.806156] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5e90cf4-9ac1-4c77-95fa-e5f7f418bf9f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.820417] env[62692]: DEBUG nova.compute.provider_tree [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 708.084020] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Releasing lock "refresh_cache-0aac5bf9-2af4-4745-a961-46986fb57779" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.084020] env[62692]: DEBUG nova.compute.manager [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 708.084020] env[62692]: DEBUG nova.compute.manager [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 708.084020] env[62692]: DEBUG nova.network.neutron [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 708.126259] env[62692]: DEBUG nova.network.neutron [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 708.281674] env[62692]: DEBUG nova.compute.manager [req-8b036b02-3699-4daf-844f-208c02b8a890 req-b2d32b4a-99a8-44bb-879f-4e6e08081d84 service nova] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Received event network-changed-eed49ef6-5628-4bd8-a95e-05d1f3c803cc {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 708.281890] env[62692]: DEBUG nova.compute.manager [req-8b036b02-3699-4daf-844f-208c02b8a890 req-b2d32b4a-99a8-44bb-879f-4e6e08081d84 service nova] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Refreshing instance network info cache due to event network-changed-eed49ef6-5628-4bd8-a95e-05d1f3c803cc. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 708.282419] env[62692]: DEBUG oslo_concurrency.lockutils [req-8b036b02-3699-4daf-844f-208c02b8a890 req-b2d32b4a-99a8-44bb-879f-4e6e08081d84 service nova] Acquiring lock "refresh_cache-2feffd08-f011-4117-9f8d-ac6730f893f5" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.282675] env[62692]: DEBUG oslo_concurrency.lockutils [req-8b036b02-3699-4daf-844f-208c02b8a890 req-b2d32b4a-99a8-44bb-879f-4e6e08081d84 service nova] Acquired lock "refresh_cache-2feffd08-f011-4117-9f8d-ac6730f893f5" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.282879] env[62692]: DEBUG nova.network.neutron [req-8b036b02-3699-4daf-844f-208c02b8a890 req-b2d32b4a-99a8-44bb-879f-4e6e08081d84 service nova] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Refreshing network info cache for port eed49ef6-5628-4bd8-a95e-05d1f3c803cc {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 708.326309] env[62692]: DEBUG nova.scheduler.client.report [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 708.462452] env[62692]: ERROR nova.compute.manager [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port eed49ef6-5628-4bd8-a95e-05d1f3c803cc, please check neutron logs for more information. [ 708.462452] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 708.462452] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 708.462452] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 708.462452] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 708.462452] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 708.462452] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 708.462452] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 708.462452] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.462452] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 708.462452] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.462452] env[62692]: ERROR nova.compute.manager raise self.value [ 708.462452] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 708.462452] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 708.462452] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.462452] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 708.462904] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.462904] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 708.462904] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port eed49ef6-5628-4bd8-a95e-05d1f3c803cc, please check neutron logs for more information. [ 708.462904] env[62692]: ERROR nova.compute.manager [ 708.462904] env[62692]: Traceback (most recent call last): [ 708.462904] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 708.462904] env[62692]: listener.cb(fileno) [ 708.462904] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 708.462904] env[62692]: result = function(*args, **kwargs) [ 708.462904] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 708.462904] env[62692]: return func(*args, **kwargs) [ 708.462904] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 708.462904] env[62692]: raise e [ 708.462904] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 708.462904] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 708.462904] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 708.462904] env[62692]: created_port_ids = self._update_ports_for_instance( [ 708.462904] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 708.462904] env[62692]: with excutils.save_and_reraise_exception(): [ 708.462904] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.462904] env[62692]: self.force_reraise() [ 708.462904] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.462904] env[62692]: raise self.value [ 708.462904] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 708.462904] env[62692]: updated_port = self._update_port( [ 708.462904] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.462904] env[62692]: _ensure_no_port_binding_failure(port) [ 708.462904] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.462904] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 708.463542] env[62692]: nova.exception.PortBindingFailed: Binding failed for port eed49ef6-5628-4bd8-a95e-05d1f3c803cc, please check neutron logs for more information. [ 708.463542] env[62692]: Removing descriptor: 19 [ 708.463542] env[62692]: ERROR nova.compute.manager [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port eed49ef6-5628-4bd8-a95e-05d1f3c803cc, please check neutron logs for more information. [ 708.463542] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Traceback (most recent call last): [ 708.463542] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 708.463542] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] yield resources [ 708.463542] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 708.463542] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] self.driver.spawn(context, instance, image_meta, [ 708.463542] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 708.463542] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 708.463542] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 708.463542] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] vm_ref = self.build_virtual_machine(instance, [ 708.463883] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 708.463883] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] vif_infos = vmwarevif.get_vif_info(self._session, [ 708.463883] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 708.463883] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] for vif in network_info: [ 708.463883] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 708.463883] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] return self._sync_wrapper(fn, *args, **kwargs) [ 708.463883] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 708.463883] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] self.wait() [ 708.463883] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 708.463883] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] self[:] = self._gt.wait() [ 708.463883] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 708.463883] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] return self._exit_event.wait() [ 708.463883] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 708.464381] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] result = hub.switch() [ 708.464381] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 708.464381] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] return self.greenlet.switch() [ 708.464381] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 708.464381] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] result = function(*args, **kwargs) [ 708.464381] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 708.464381] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] return func(*args, **kwargs) [ 708.464381] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 708.464381] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] raise e [ 708.464381] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 708.464381] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] nwinfo = self.network_api.allocate_for_instance( [ 708.464381] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 708.464381] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] created_port_ids = self._update_ports_for_instance( [ 708.464748] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 708.464748] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] with excutils.save_and_reraise_exception(): [ 708.464748] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.464748] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] self.force_reraise() [ 708.464748] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.464748] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] raise self.value [ 708.464748] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 708.464748] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] updated_port = self._update_port( [ 708.464748] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.464748] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] _ensure_no_port_binding_failure(port) [ 708.464748] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.464748] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] raise exception.PortBindingFailed(port_id=port['id']) [ 708.465047] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] nova.exception.PortBindingFailed: Binding failed for port eed49ef6-5628-4bd8-a95e-05d1f3c803cc, please check neutron logs for more information. [ 708.465047] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] [ 708.465047] env[62692]: INFO nova.compute.manager [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Terminating instance [ 708.467213] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Acquiring lock "refresh_cache-2feffd08-f011-4117-9f8d-ac6730f893f5" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.635095] env[62692]: DEBUG nova.network.neutron [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.815676] env[62692]: DEBUG nova.network.neutron [req-8b036b02-3699-4daf-844f-208c02b8a890 req-b2d32b4a-99a8-44bb-879f-4e6e08081d84 service nova] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 708.836257] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.130s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.837033] env[62692]: ERROR nova.compute.manager [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 501c0ab4-2706-44d5-9057-d5ae4b34ae1f, please check neutron logs for more information. [ 708.837033] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Traceback (most recent call last): [ 708.837033] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 708.837033] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] self.driver.spawn(context, instance, image_meta, [ 708.837033] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 708.837033] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 708.837033] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 708.837033] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] vm_ref = self.build_virtual_machine(instance, [ 708.837033] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 708.837033] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] vif_infos = vmwarevif.get_vif_info(self._session, [ 708.837033] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 708.837606] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] for vif in network_info: [ 708.837606] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 708.837606] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] return self._sync_wrapper(fn, *args, **kwargs) [ 708.837606] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 708.837606] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] self.wait() [ 708.837606] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 708.837606] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] self[:] = self._gt.wait() [ 708.837606] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 708.837606] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] return self._exit_event.wait() [ 708.837606] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 708.837606] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] result = hub.switch() [ 708.837606] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 708.837606] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] return self.greenlet.switch() [ 708.838213] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 708.838213] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] result = function(*args, **kwargs) [ 708.838213] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 708.838213] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] return func(*args, **kwargs) [ 708.838213] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 708.838213] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] raise e [ 708.838213] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 708.838213] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] nwinfo = self.network_api.allocate_for_instance( [ 708.838213] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 708.838213] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] created_port_ids = self._update_ports_for_instance( [ 708.838213] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 708.838213] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] with excutils.save_and_reraise_exception(): [ 708.838213] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.838758] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] self.force_reraise() [ 708.838758] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.838758] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] raise self.value [ 708.838758] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 708.838758] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] updated_port = self._update_port( [ 708.838758] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.838758] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] _ensure_no_port_binding_failure(port) [ 708.838758] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.838758] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] raise exception.PortBindingFailed(port_id=port['id']) [ 708.838758] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] nova.exception.PortBindingFailed: Binding failed for port 501c0ab4-2706-44d5-9057-d5ae4b34ae1f, please check neutron logs for more information. [ 708.838758] env[62692]: ERROR nova.compute.manager [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] [ 708.839270] env[62692]: DEBUG nova.compute.utils [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Binding failed for port 501c0ab4-2706-44d5-9057-d5ae4b34ae1f, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 708.839270] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.473s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.843688] env[62692]: INFO nova.compute.claims [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 708.845971] env[62692]: DEBUG nova.compute.manager [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Build of instance 5d18922d-5937-47c8-a04a-aa757c153e8f was re-scheduled: Binding failed for port 501c0ab4-2706-44d5-9057-d5ae4b34ae1f, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 708.847645] env[62692]: DEBUG nova.compute.manager [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 708.847645] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Acquiring lock "refresh_cache-5d18922d-5937-47c8-a04a-aa757c153e8f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.847645] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Acquired lock "refresh_cache-5d18922d-5937-47c8-a04a-aa757c153e8f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.847645] env[62692]: DEBUG nova.network.neutron [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 708.926508] env[62692]: DEBUG nova.network.neutron [req-8b036b02-3699-4daf-844f-208c02b8a890 req-b2d32b4a-99a8-44bb-879f-4e6e08081d84 service nova] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.953605] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "b1d6409a-2733-470f-a929-672fe1631b1b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.953979] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "b1d6409a-2733-470f-a929-672fe1631b1b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.619516] env[62692]: INFO nova.compute.manager [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] [instance: 0aac5bf9-2af4-4745-a961-46986fb57779] Took 1.54 seconds to deallocate network for instance. [ 709.626639] env[62692]: DEBUG oslo_concurrency.lockutils [req-8b036b02-3699-4daf-844f-208c02b8a890 req-b2d32b4a-99a8-44bb-879f-4e6e08081d84 service nova] Releasing lock "refresh_cache-2feffd08-f011-4117-9f8d-ac6730f893f5" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.627567] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Acquired lock "refresh_cache-2feffd08-f011-4117-9f8d-ac6730f893f5" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.627753] env[62692]: DEBUG nova.network.neutron [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 709.652444] env[62692]: DEBUG nova.network.neutron [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 709.800922] env[62692]: DEBUG nova.network.neutron [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.160921] env[62692]: DEBUG nova.network.neutron [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 710.296566] env[62692]: DEBUG nova.network.neutron [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.305111] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Releasing lock "refresh_cache-5d18922d-5937-47c8-a04a-aa757c153e8f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.305111] env[62692]: DEBUG nova.compute.manager [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 710.305237] env[62692]: DEBUG nova.compute.manager [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 710.305373] env[62692]: DEBUG nova.network.neutron [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 710.317915] env[62692]: DEBUG nova.compute.manager [req-dac96672-3b23-4eb3-96c2-dc645f885cbc req-10e0fcbb-4e27-4b97-8c4c-2810c8a83da8 service nova] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Received event network-vif-deleted-eed49ef6-5628-4bd8-a95e-05d1f3c803cc {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 710.333197] env[62692]: DEBUG nova.network.neutron [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 710.565326] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2b8a84b-f242-4103-a3df-d1b918837ef8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.573021] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f860f5d-43f2-41ed-86f8-1549161ddc20 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.608767] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-210449bd-5295-4032-a8ea-8700d31133e2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.616234] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86dde0f3-5eaf-47bc-a68c-3bfe8227c7cd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.630359] env[62692]: DEBUG nova.compute.provider_tree [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 710.656457] env[62692]: INFO nova.scheduler.client.report [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Deleted allocations for instance 0aac5bf9-2af4-4745-a961-46986fb57779 [ 710.802123] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Releasing lock "refresh_cache-2feffd08-f011-4117-9f8d-ac6730f893f5" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.802593] env[62692]: DEBUG nova.compute.manager [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 710.802790] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 710.803106] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bc7befa9-1d3a-4a01-9276-4428dbbe085f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.813414] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a5c2f77-d3ab-4025-bb36-f33ad29e6945 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.834988] env[62692]: DEBUG nova.network.neutron [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.853639] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2feffd08-f011-4117-9f8d-ac6730f893f5 could not be found. [ 710.854105] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 710.854364] env[62692]: INFO nova.compute.manager [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Took 0.05 seconds to destroy the instance on the hypervisor. [ 710.854664] env[62692]: DEBUG oslo.service.loopingcall [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 710.856022] env[62692]: DEBUG nova.compute.manager [-] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 710.856120] env[62692]: DEBUG nova.network.neutron [-] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 710.883846] env[62692]: DEBUG nova.network.neutron [-] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 711.132783] env[62692]: DEBUG nova.scheduler.client.report [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 711.161556] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9b023088-8030-4542-9b6d-5e23f0d4ddd1 tempest-ListImageFiltersTestJSON-1280724270 tempest-ListImageFiltersTestJSON-1280724270-project-member] Lock "0aac5bf9-2af4-4745-a961-46986fb57779" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.226s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.337696] env[62692]: INFO nova.compute.manager [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] [instance: 5d18922d-5937-47c8-a04a-aa757c153e8f] Took 1.03 seconds to deallocate network for instance. [ 711.386107] env[62692]: DEBUG nova.network.neutron [-] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.638259] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.799s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.638828] env[62692]: DEBUG nova.compute.manager [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 711.642171] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.089s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.644552] env[62692]: INFO nova.compute.claims [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 711.664658] env[62692]: DEBUG nova.compute.manager [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 711.889463] env[62692]: INFO nova.compute.manager [-] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Took 1.03 seconds to deallocate network for instance. [ 711.892185] env[62692]: DEBUG nova.compute.claims [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 711.892423] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.151508] env[62692]: DEBUG nova.compute.utils [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 712.152915] env[62692]: DEBUG nova.compute.manager [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 712.153111] env[62692]: DEBUG nova.network.neutron [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 712.192512] env[62692]: DEBUG oslo_concurrency.lockutils [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.231197] env[62692]: DEBUG nova.policy [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b52c2aa90394303a127d9ece53cd78f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa15800872af4c2393abee0ec1e43d59', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 712.381043] env[62692]: INFO nova.scheduler.client.report [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Deleted allocations for instance 5d18922d-5937-47c8-a04a-aa757c153e8f [ 712.661241] env[62692]: DEBUG nova.compute.manager [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 712.795316] env[62692]: DEBUG nova.network.neutron [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Successfully created port: c6d0b984-14e0-4bea-b8d5-82c7df1084b3 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 712.806343] env[62692]: DEBUG oslo_concurrency.lockutils [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Acquiring lock "ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.807090] env[62692]: DEBUG oslo_concurrency.lockutils [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Lock "ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.891098] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4a5f9aa4-9a17-4f71-a523-8342ae186462 tempest-ServersV294TestFqdnHostnames-1109156262 tempest-ServersV294TestFqdnHostnames-1109156262-project-member] Lock "5d18922d-5937-47c8-a04a-aa757c153e8f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.903s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.158424] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ac57a70-1e60-4e83-9dc1-dded6a6052ff {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.168073] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3930d142-dab9-437a-952d-d4b38ceb61dd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.210080] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c4ecb48-132a-4ed1-9839-9f12f1608ecc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.221215] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bec8ed30-a917-4fe9-ae3f-a18a2a786598 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.236277] env[62692]: DEBUG nova.compute.provider_tree [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 713.392876] env[62692]: DEBUG nova.compute.manager [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 713.678764] env[62692]: DEBUG nova.compute.manager [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 713.714698] env[62692]: DEBUG nova.virt.hardware [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 713.715073] env[62692]: DEBUG nova.virt.hardware [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 713.715210] env[62692]: DEBUG nova.virt.hardware [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 713.715412] env[62692]: DEBUG nova.virt.hardware [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 713.716128] env[62692]: DEBUG nova.virt.hardware [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 713.716128] env[62692]: DEBUG nova.virt.hardware [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 713.716128] env[62692]: DEBUG nova.virt.hardware [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 713.716128] env[62692]: DEBUG nova.virt.hardware [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 713.716295] env[62692]: DEBUG nova.virt.hardware [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 713.716456] env[62692]: DEBUG nova.virt.hardware [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 713.716628] env[62692]: DEBUG nova.virt.hardware [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 713.717490] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68552538-ee99-4ef5-a64a-282ce0112b00 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.726399] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e8ed62e-1e1e-456b-9f0f-df00c023b063 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.748903] env[62692]: DEBUG nova.scheduler.client.report [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 713.821175] env[62692]: DEBUG nova.compute.manager [req-22ff00ea-ed8c-4954-9788-7178401084ae req-5b26850f-9126-4dde-b240-5e2e18eb6ef4 service nova] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Received event network-changed-c6d0b984-14e0-4bea-b8d5-82c7df1084b3 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 713.821811] env[62692]: DEBUG nova.compute.manager [req-22ff00ea-ed8c-4954-9788-7178401084ae req-5b26850f-9126-4dde-b240-5e2e18eb6ef4 service nova] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Refreshing instance network info cache due to event network-changed-c6d0b984-14e0-4bea-b8d5-82c7df1084b3. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 713.821811] env[62692]: DEBUG oslo_concurrency.lockutils [req-22ff00ea-ed8c-4954-9788-7178401084ae req-5b26850f-9126-4dde-b240-5e2e18eb6ef4 service nova] Acquiring lock "refresh_cache-9422a9a8-795d-4794-8ba9-9e6f88cdb464" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.821811] env[62692]: DEBUG oslo_concurrency.lockutils [req-22ff00ea-ed8c-4954-9788-7178401084ae req-5b26850f-9126-4dde-b240-5e2e18eb6ef4 service nova] Acquired lock "refresh_cache-9422a9a8-795d-4794-8ba9-9e6f88cdb464" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.821995] env[62692]: DEBUG nova.network.neutron [req-22ff00ea-ed8c-4954-9788-7178401084ae req-5b26850f-9126-4dde-b240-5e2e18eb6ef4 service nova] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Refreshing network info cache for port c6d0b984-14e0-4bea-b8d5-82c7df1084b3 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 713.917161] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.038996] env[62692]: ERROR nova.compute.manager [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c6d0b984-14e0-4bea-b8d5-82c7df1084b3, please check neutron logs for more information. [ 714.038996] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 714.038996] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 714.038996] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 714.038996] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 714.038996] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 714.038996] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 714.038996] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 714.038996] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.038996] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 714.038996] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.038996] env[62692]: ERROR nova.compute.manager raise self.value [ 714.038996] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 714.038996] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 714.038996] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.038996] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 714.039895] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.039895] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 714.039895] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c6d0b984-14e0-4bea-b8d5-82c7df1084b3, please check neutron logs for more information. [ 714.039895] env[62692]: ERROR nova.compute.manager [ 714.039895] env[62692]: Traceback (most recent call last): [ 714.039895] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 714.039895] env[62692]: listener.cb(fileno) [ 714.039895] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 714.039895] env[62692]: result = function(*args, **kwargs) [ 714.039895] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 714.039895] env[62692]: return func(*args, **kwargs) [ 714.039895] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 714.039895] env[62692]: raise e [ 714.039895] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 714.039895] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 714.039895] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 714.039895] env[62692]: created_port_ids = self._update_ports_for_instance( [ 714.039895] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 714.039895] env[62692]: with excutils.save_and_reraise_exception(): [ 714.039895] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.039895] env[62692]: self.force_reraise() [ 714.039895] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.039895] env[62692]: raise self.value [ 714.039895] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 714.039895] env[62692]: updated_port = self._update_port( [ 714.039895] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.039895] env[62692]: _ensure_no_port_binding_failure(port) [ 714.039895] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.039895] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 714.040610] env[62692]: nova.exception.PortBindingFailed: Binding failed for port c6d0b984-14e0-4bea-b8d5-82c7df1084b3, please check neutron logs for more information. [ 714.040610] env[62692]: Removing descriptor: 19 [ 714.040610] env[62692]: ERROR nova.compute.manager [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c6d0b984-14e0-4bea-b8d5-82c7df1084b3, please check neutron logs for more information. [ 714.040610] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Traceback (most recent call last): [ 714.040610] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 714.040610] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] yield resources [ 714.040610] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 714.040610] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] self.driver.spawn(context, instance, image_meta, [ 714.040610] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 714.040610] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] self._vmops.spawn(context, instance, image_meta, injected_files, [ 714.040610] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 714.040610] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] vm_ref = self.build_virtual_machine(instance, [ 714.040896] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 714.040896] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] vif_infos = vmwarevif.get_vif_info(self._session, [ 714.040896] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 714.040896] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] for vif in network_info: [ 714.040896] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 714.040896] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] return self._sync_wrapper(fn, *args, **kwargs) [ 714.040896] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 714.040896] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] self.wait() [ 714.040896] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 714.040896] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] self[:] = self._gt.wait() [ 714.040896] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 714.040896] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] return self._exit_event.wait() [ 714.040896] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 714.041227] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] result = hub.switch() [ 714.041227] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 714.041227] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] return self.greenlet.switch() [ 714.041227] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 714.041227] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] result = function(*args, **kwargs) [ 714.041227] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 714.041227] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] return func(*args, **kwargs) [ 714.041227] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 714.041227] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] raise e [ 714.041227] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 714.041227] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] nwinfo = self.network_api.allocate_for_instance( [ 714.041227] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 714.041227] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] created_port_ids = self._update_ports_for_instance( [ 714.041530] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 714.041530] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] with excutils.save_and_reraise_exception(): [ 714.041530] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.041530] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] self.force_reraise() [ 714.041530] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.041530] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] raise self.value [ 714.041530] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 714.041530] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] updated_port = self._update_port( [ 714.041530] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.041530] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] _ensure_no_port_binding_failure(port) [ 714.041530] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.041530] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] raise exception.PortBindingFailed(port_id=port['id']) [ 714.041802] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] nova.exception.PortBindingFailed: Binding failed for port c6d0b984-14e0-4bea-b8d5-82c7df1084b3, please check neutron logs for more information. [ 714.041802] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] [ 714.041802] env[62692]: INFO nova.compute.manager [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Terminating instance [ 714.043593] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Acquiring lock "refresh_cache-9422a9a8-795d-4794-8ba9-9e6f88cdb464" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.255312] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.613s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.255910] env[62692]: DEBUG nova.compute.manager [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 714.262595] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.157s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.352401] env[62692]: DEBUG nova.network.neutron [req-22ff00ea-ed8c-4954-9788-7178401084ae req-5b26850f-9126-4dde-b240-5e2e18eb6ef4 service nova] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 714.425655] env[62692]: DEBUG nova.network.neutron [req-22ff00ea-ed8c-4954-9788-7178401084ae req-5b26850f-9126-4dde-b240-5e2e18eb6ef4 service nova] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.771335] env[62692]: DEBUG nova.compute.utils [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 714.777067] env[62692]: DEBUG nova.compute.manager [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 714.777067] env[62692]: DEBUG nova.network.neutron [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 714.837492] env[62692]: DEBUG nova.policy [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd33baeebac3d4fad907677f763bf71be', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '43057671f72b4bfa83f33855a1bf4422', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 714.928566] env[62692]: DEBUG oslo_concurrency.lockutils [req-22ff00ea-ed8c-4954-9788-7178401084ae req-5b26850f-9126-4dde-b240-5e2e18eb6ef4 service nova] Releasing lock "refresh_cache-9422a9a8-795d-4794-8ba9-9e6f88cdb464" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.928983] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Acquired lock "refresh_cache-9422a9a8-795d-4794-8ba9-9e6f88cdb464" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.929203] env[62692]: DEBUG nova.network.neutron [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 715.241041] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f7af2e6-e107-4872-82be-b0dd75c44dea {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.251465] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd4b12b1-a03f-44fd-87b3-0b27ae5e5668 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.294267] env[62692]: DEBUG nova.compute.manager [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 715.298087] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cac5cb9-6d65-4e8a-abf6-3a69d2fb2158 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.307145] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae5b0df7-97c3-46ce-9bc6-055e01264216 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.324105] env[62692]: DEBUG nova.compute.provider_tree [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 715.337584] env[62692]: DEBUG nova.network.neutron [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Successfully created port: 1c476cc3-b128-4738-a1d2-8d915dfee19b {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 715.461626] env[62692]: DEBUG nova.network.neutron [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 715.602585] env[62692]: DEBUG nova.network.neutron [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.828300] env[62692]: DEBUG nova.scheduler.client.report [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 715.870493] env[62692]: DEBUG nova.compute.manager [req-d5d635bd-8efb-4b45-8777-4826c2e1cf67 req-023addf0-e506-4abb-9b4c-7e375b1dfc99 service nova] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Received event network-vif-deleted-c6d0b984-14e0-4bea-b8d5-82c7df1084b3 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 716.103745] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Releasing lock "refresh_cache-9422a9a8-795d-4794-8ba9-9e6f88cdb464" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.104215] env[62692]: DEBUG nova.compute.manager [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 716.104468] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 716.104716] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cd0bac53-9ec5-4ba0-917b-4810d10626ab {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.113607] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad9fd15-2954-40d0-9b84-cf827f8c027d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.137264] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9422a9a8-795d-4794-8ba9-9e6f88cdb464 could not be found. [ 716.137547] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 716.137767] env[62692]: INFO nova.compute.manager [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Took 0.03 seconds to destroy the instance on the hypervisor. [ 716.138125] env[62692]: DEBUG oslo.service.loopingcall [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 716.138372] env[62692]: DEBUG nova.compute.manager [-] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 716.138480] env[62692]: DEBUG nova.network.neutron [-] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 716.156607] env[62692]: DEBUG nova.network.neutron [-] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 716.307333] env[62692]: DEBUG nova.compute.manager [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 716.335572] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.073s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.335708] env[62692]: ERROR nova.compute.manager [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 451c23dc-3dfb-4dc8-9e9b-eb70351b81bb, please check neutron logs for more information. [ 716.335708] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Traceback (most recent call last): [ 716.335708] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 716.335708] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] self.driver.spawn(context, instance, image_meta, [ 716.335708] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 716.335708] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 716.335708] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 716.335708] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] vm_ref = self.build_virtual_machine(instance, [ 716.335708] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 716.335708] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] vif_infos = vmwarevif.get_vif_info(self._session, [ 716.335708] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 716.336063] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] for vif in network_info: [ 716.336063] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 716.336063] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] return self._sync_wrapper(fn, *args, **kwargs) [ 716.336063] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 716.336063] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] self.wait() [ 716.336063] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 716.336063] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] self[:] = self._gt.wait() [ 716.336063] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 716.336063] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] return self._exit_event.wait() [ 716.336063] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 716.336063] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] result = hub.switch() [ 716.336063] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 716.336063] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] return self.greenlet.switch() [ 716.336429] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 716.336429] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] result = function(*args, **kwargs) [ 716.336429] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 716.336429] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] return func(*args, **kwargs) [ 716.336429] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 716.336429] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] raise e [ 716.336429] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.336429] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] nwinfo = self.network_api.allocate_for_instance( [ 716.336429] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 716.336429] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] created_port_ids = self._update_ports_for_instance( [ 716.336429] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 716.336429] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] with excutils.save_and_reraise_exception(): [ 716.336429] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.336792] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] self.force_reraise() [ 716.336792] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.336792] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] raise self.value [ 716.336792] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 716.336792] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] updated_port = self._update_port( [ 716.336792] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.336792] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] _ensure_no_port_binding_failure(port) [ 716.336792] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.336792] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] raise exception.PortBindingFailed(port_id=port['id']) [ 716.336792] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] nova.exception.PortBindingFailed: Binding failed for port 451c23dc-3dfb-4dc8-9e9b-eb70351b81bb, please check neutron logs for more information. [ 716.336792] env[62692]: ERROR nova.compute.manager [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] [ 716.337096] env[62692]: DEBUG nova.compute.utils [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Binding failed for port 451c23dc-3dfb-4dc8-9e9b-eb70351b81bb, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 716.341770] env[62692]: DEBUG nova.virt.hardware [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 716.343419] env[62692]: DEBUG nova.virt.hardware [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 716.343419] env[62692]: DEBUG nova.virt.hardware [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 716.343419] env[62692]: DEBUG nova.virt.hardware [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 716.343419] env[62692]: DEBUG nova.virt.hardware [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 716.343419] env[62692]: DEBUG nova.virt.hardware [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 716.343612] env[62692]: DEBUG nova.virt.hardware [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 716.343612] env[62692]: DEBUG nova.virt.hardware [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 716.343612] env[62692]: DEBUG nova.virt.hardware [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 716.343612] env[62692]: DEBUG nova.virt.hardware [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 716.343612] env[62692]: DEBUG nova.virt.hardware [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 716.343772] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.036s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.345178] env[62692]: INFO nova.compute.claims [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 716.349019] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00977ee7-ec6f-4c47-b28a-e21910b48909 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.351254] env[62692]: DEBUG nova.compute.manager [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Build of instance b76cce5f-b4ab-424f-b604-684f5c22154a was re-scheduled: Binding failed for port 451c23dc-3dfb-4dc8-9e9b-eb70351b81bb, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 716.351669] env[62692]: DEBUG nova.compute.manager [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 716.351889] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "refresh_cache-b76cce5f-b4ab-424f-b604-684f5c22154a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.352050] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquired lock "refresh_cache-b76cce5f-b4ab-424f-b604-684f5c22154a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.352211] env[62692]: DEBUG nova.network.neutron [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 716.359178] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6321af27-0fb4-4e22-9e09-6dfe18f11ff6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.424640] env[62692]: ERROR nova.compute.manager [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1c476cc3-b128-4738-a1d2-8d915dfee19b, please check neutron logs for more information. [ 716.424640] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 716.424640] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.424640] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 716.424640] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 716.424640] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 716.424640] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 716.424640] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 716.424640] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.424640] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 716.424640] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.424640] env[62692]: ERROR nova.compute.manager raise self.value [ 716.424640] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 716.424640] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 716.424640] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.424640] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 716.425067] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.425067] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 716.425067] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1c476cc3-b128-4738-a1d2-8d915dfee19b, please check neutron logs for more information. [ 716.425067] env[62692]: ERROR nova.compute.manager [ 716.425067] env[62692]: Traceback (most recent call last): [ 716.425067] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 716.425067] env[62692]: listener.cb(fileno) [ 716.425067] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 716.425067] env[62692]: result = function(*args, **kwargs) [ 716.425067] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 716.425067] env[62692]: return func(*args, **kwargs) [ 716.425067] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 716.425067] env[62692]: raise e [ 716.425067] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.425067] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 716.425067] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 716.425067] env[62692]: created_port_ids = self._update_ports_for_instance( [ 716.425067] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 716.425067] env[62692]: with excutils.save_and_reraise_exception(): [ 716.425067] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.425067] env[62692]: self.force_reraise() [ 716.425067] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.425067] env[62692]: raise self.value [ 716.425067] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 716.425067] env[62692]: updated_port = self._update_port( [ 716.425067] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.425067] env[62692]: _ensure_no_port_binding_failure(port) [ 716.425067] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.425067] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 716.425926] env[62692]: nova.exception.PortBindingFailed: Binding failed for port 1c476cc3-b128-4738-a1d2-8d915dfee19b, please check neutron logs for more information. [ 716.425926] env[62692]: Removing descriptor: 19 [ 716.425926] env[62692]: ERROR nova.compute.manager [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1c476cc3-b128-4738-a1d2-8d915dfee19b, please check neutron logs for more information. [ 716.425926] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Traceback (most recent call last): [ 716.425926] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 716.425926] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] yield resources [ 716.425926] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 716.425926] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] self.driver.spawn(context, instance, image_meta, [ 716.425926] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 716.425926] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] self._vmops.spawn(context, instance, image_meta, injected_files, [ 716.425926] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 716.425926] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] vm_ref = self.build_virtual_machine(instance, [ 716.426248] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 716.426248] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] vif_infos = vmwarevif.get_vif_info(self._session, [ 716.426248] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 716.426248] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] for vif in network_info: [ 716.426248] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 716.426248] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] return self._sync_wrapper(fn, *args, **kwargs) [ 716.426248] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 716.426248] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] self.wait() [ 716.426248] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 716.426248] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] self[:] = self._gt.wait() [ 716.426248] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 716.426248] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] return self._exit_event.wait() [ 716.426248] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 716.426555] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] result = hub.switch() [ 716.426555] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 716.426555] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] return self.greenlet.switch() [ 716.426555] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 716.426555] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] result = function(*args, **kwargs) [ 716.426555] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 716.426555] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] return func(*args, **kwargs) [ 716.426555] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 716.426555] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] raise e [ 716.426555] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.426555] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] nwinfo = self.network_api.allocate_for_instance( [ 716.426555] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 716.426555] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] created_port_ids = self._update_ports_for_instance( [ 716.426929] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 716.426929] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] with excutils.save_and_reraise_exception(): [ 716.426929] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.426929] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] self.force_reraise() [ 716.426929] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.426929] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] raise self.value [ 716.426929] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 716.426929] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] updated_port = self._update_port( [ 716.426929] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.426929] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] _ensure_no_port_binding_failure(port) [ 716.426929] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.426929] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] raise exception.PortBindingFailed(port_id=port['id']) [ 716.427759] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] nova.exception.PortBindingFailed: Binding failed for port 1c476cc3-b128-4738-a1d2-8d915dfee19b, please check neutron logs for more information. [ 716.427759] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] [ 716.427759] env[62692]: INFO nova.compute.manager [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Terminating instance [ 716.429031] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquiring lock "refresh_cache-146e164e-2543-46ab-a7d3-9803d56f61ec" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.429218] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquired lock "refresh_cache-146e164e-2543-46ab-a7d3-9803d56f61ec" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.429387] env[62692]: DEBUG nova.network.neutron [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 716.660165] env[62692]: DEBUG nova.network.neutron [-] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.872060] env[62692]: DEBUG nova.network.neutron [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 716.950824] env[62692]: DEBUG nova.network.neutron [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 716.995277] env[62692]: DEBUG nova.network.neutron [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.026666] env[62692]: DEBUG nova.network.neutron [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.163596] env[62692]: INFO nova.compute.manager [-] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Took 1.02 seconds to deallocate network for instance. [ 717.166418] env[62692]: DEBUG nova.compute.claims [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 717.166598] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.501685] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Releasing lock "refresh_cache-b76cce5f-b4ab-424f-b604-684f5c22154a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.501941] env[62692]: DEBUG nova.compute.manager [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 717.502161] env[62692]: DEBUG nova.compute.manager [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 717.502339] env[62692]: DEBUG nova.network.neutron [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 717.520252] env[62692]: DEBUG nova.network.neutron [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 717.528749] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Releasing lock "refresh_cache-146e164e-2543-46ab-a7d3-9803d56f61ec" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.529317] env[62692]: DEBUG nova.compute.manager [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 717.529559] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 717.529878] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fd416a9e-782c-4aa9-bef4-b8b1a70ec59d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.544012] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8256edb-f42f-480a-87fa-6175078b4276 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.570661] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 146e164e-2543-46ab-a7d3-9803d56f61ec could not be found. [ 717.570661] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 717.570661] env[62692]: INFO nova.compute.manager [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Took 0.04 seconds to destroy the instance on the hypervisor. [ 717.570661] env[62692]: DEBUG oslo.service.loopingcall [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 717.572714] env[62692]: DEBUG nova.compute.manager [-] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 717.572954] env[62692]: DEBUG nova.network.neutron [-] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 717.588043] env[62692]: DEBUG nova.network.neutron [-] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 717.724143] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85e9d108-eb44-49e5-8642-7edb7d527bc7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.732130] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f0aa006-b8e4-4cff-9a10-c4d896bbd9b8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.762624] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33e68230-5a4e-414f-98ef-c86a4e4b9ef6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.770288] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21fb7a62-3fa4-48b0-bcf4-075ff473ada1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.782740] env[62692]: DEBUG nova.compute.provider_tree [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.908309] env[62692]: DEBUG nova.compute.manager [req-f9b86021-0962-43a4-bca7-030bb118cb0f req-7455a789-9072-4be3-a0e6-62d03742ea9b service nova] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Received event network-changed-1c476cc3-b128-4738-a1d2-8d915dfee19b {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 717.908554] env[62692]: DEBUG nova.compute.manager [req-f9b86021-0962-43a4-bca7-030bb118cb0f req-7455a789-9072-4be3-a0e6-62d03742ea9b service nova] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Refreshing instance network info cache due to event network-changed-1c476cc3-b128-4738-a1d2-8d915dfee19b. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 717.908709] env[62692]: DEBUG oslo_concurrency.lockutils [req-f9b86021-0962-43a4-bca7-030bb118cb0f req-7455a789-9072-4be3-a0e6-62d03742ea9b service nova] Acquiring lock "refresh_cache-146e164e-2543-46ab-a7d3-9803d56f61ec" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.908851] env[62692]: DEBUG oslo_concurrency.lockutils [req-f9b86021-0962-43a4-bca7-030bb118cb0f req-7455a789-9072-4be3-a0e6-62d03742ea9b service nova] Acquired lock "refresh_cache-146e164e-2543-46ab-a7d3-9803d56f61ec" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.909018] env[62692]: DEBUG nova.network.neutron [req-f9b86021-0962-43a4-bca7-030bb118cb0f req-7455a789-9072-4be3-a0e6-62d03742ea9b service nova] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Refreshing network info cache for port 1c476cc3-b128-4738-a1d2-8d915dfee19b {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 718.023096] env[62692]: DEBUG nova.network.neutron [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.089904] env[62692]: DEBUG nova.network.neutron [-] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.287765] env[62692]: DEBUG nova.scheduler.client.report [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 718.427734] env[62692]: DEBUG nova.network.neutron [req-f9b86021-0962-43a4-bca7-030bb118cb0f req-7455a789-9072-4be3-a0e6-62d03742ea9b service nova] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 718.489109] env[62692]: DEBUG nova.network.neutron [req-f9b86021-0962-43a4-bca7-030bb118cb0f req-7455a789-9072-4be3-a0e6-62d03742ea9b service nova] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.525538] env[62692]: INFO nova.compute.manager [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b76cce5f-b4ab-424f-b604-684f5c22154a] Took 1.02 seconds to deallocate network for instance. [ 718.592517] env[62692]: INFO nova.compute.manager [-] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Took 1.02 seconds to deallocate network for instance. [ 718.594850] env[62692]: DEBUG nova.compute.claims [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 718.595066] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.751726] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 718.751726] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 718.794318] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.450s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.794883] env[62692]: DEBUG nova.compute.manager [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 718.797451] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.395s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.798976] env[62692]: INFO nova.compute.claims [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 718.992106] env[62692]: DEBUG oslo_concurrency.lockutils [req-f9b86021-0962-43a4-bca7-030bb118cb0f req-7455a789-9072-4be3-a0e6-62d03742ea9b service nova] Releasing lock "refresh_cache-146e164e-2543-46ab-a7d3-9803d56f61ec" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.992106] env[62692]: DEBUG nova.compute.manager [req-f9b86021-0962-43a4-bca7-030bb118cb0f req-7455a789-9072-4be3-a0e6-62d03742ea9b service nova] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Received event network-vif-deleted-1c476cc3-b128-4738-a1d2-8d915dfee19b {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 719.258113] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 719.258113] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Starting heal instance info cache {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 719.258113] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Rebuilding the list of instances to heal {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 719.303220] env[62692]: DEBUG nova.compute.utils [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 719.306711] env[62692]: DEBUG nova.compute.manager [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 719.306942] env[62692]: DEBUG nova.network.neutron [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 719.353162] env[62692]: DEBUG nova.policy [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '249a3b566fcf4d88a973693512df370e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0b0b2e09a06a4fc7a00a03f72989ffed', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 719.567903] env[62692]: INFO nova.scheduler.client.report [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Deleted allocations for instance b76cce5f-b4ab-424f-b604-684f5c22154a [ 719.686338] env[62692]: DEBUG nova.network.neutron [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Successfully created port: e8165560-d358-4cf0-815e-2e3555f92c1f {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 719.761111] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Skipping network cache update for instance because it is Building. {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 719.761293] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Skipping network cache update for instance because it is Building. {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 719.761429] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Skipping network cache update for instance because it is Building. {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 719.761557] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Skipping network cache update for instance because it is Building. {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 719.761693] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Skipping network cache update for instance because it is Building. {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 719.761820] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Didn't find any instances for network info cache update. {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 719.762037] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 719.762199] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 719.762343] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 719.762489] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 719.762626] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 719.762768] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 719.762962] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62692) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 719.763451] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager.update_available_resource {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 719.807669] env[62692]: DEBUG nova.compute.manager [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 720.076231] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd5c67ad-c485-419d-a980-cfd22512441b tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "b76cce5f-b4ab-424f-b604-684f5c22154a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.450s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.266612] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.272050] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3992556c-983a-4791-af51-6a0677942022 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.280419] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e91a968-8125-4259-816d-a1ee6637eeb7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.310219] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5b7477d-dd99-4f7b-905b-f05e90c6f35f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.320373] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7683e34-2f91-4fcb-ab35-79d8a5c7bac9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.334714] env[62692]: DEBUG nova.compute.provider_tree [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.581858] env[62692]: DEBUG nova.compute.manager [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 720.645151] env[62692]: DEBUG nova.compute.manager [req-59f0f69b-d2d6-4c2d-ad25-09abd30a9a59 req-5f69a6d0-ac68-4136-bba4-f0757cd337f4 service nova] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Received event network-changed-e8165560-d358-4cf0-815e-2e3555f92c1f {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 720.645151] env[62692]: DEBUG nova.compute.manager [req-59f0f69b-d2d6-4c2d-ad25-09abd30a9a59 req-5f69a6d0-ac68-4136-bba4-f0757cd337f4 service nova] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Refreshing instance network info cache due to event network-changed-e8165560-d358-4cf0-815e-2e3555f92c1f. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 720.645151] env[62692]: DEBUG oslo_concurrency.lockutils [req-59f0f69b-d2d6-4c2d-ad25-09abd30a9a59 req-5f69a6d0-ac68-4136-bba4-f0757cd337f4 service nova] Acquiring lock "refresh_cache-dbd51115-6b41-4019-955d-16a52b534e84" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.645151] env[62692]: DEBUG oslo_concurrency.lockutils [req-59f0f69b-d2d6-4c2d-ad25-09abd30a9a59 req-5f69a6d0-ac68-4136-bba4-f0757cd337f4 service nova] Acquired lock "refresh_cache-dbd51115-6b41-4019-955d-16a52b534e84" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.645151] env[62692]: DEBUG nova.network.neutron [req-59f0f69b-d2d6-4c2d-ad25-09abd30a9a59 req-5f69a6d0-ac68-4136-bba4-f0757cd337f4 service nova] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Refreshing network info cache for port e8165560-d358-4cf0-815e-2e3555f92c1f {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 720.823039] env[62692]: DEBUG nova.compute.manager [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 720.837725] env[62692]: DEBUG nova.scheduler.client.report [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 720.851544] env[62692]: DEBUG nova.virt.hardware [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 720.852300] env[62692]: DEBUG nova.virt.hardware [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 720.852639] env[62692]: DEBUG nova.virt.hardware [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 720.852966] env[62692]: DEBUG nova.virt.hardware [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 720.853752] env[62692]: DEBUG nova.virt.hardware [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 720.853752] env[62692]: DEBUG nova.virt.hardware [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 720.853752] env[62692]: DEBUG nova.virt.hardware [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 720.853752] env[62692]: DEBUG nova.virt.hardware [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 720.853974] env[62692]: DEBUG nova.virt.hardware [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 720.854212] env[62692]: DEBUG nova.virt.hardware [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 720.854441] env[62692]: DEBUG nova.virt.hardware [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 720.855339] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3d0860a-e2ef-4680-886c-6a7a9528bcdf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.865190] env[62692]: ERROR nova.compute.manager [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e8165560-d358-4cf0-815e-2e3555f92c1f, please check neutron logs for more information. [ 720.865190] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 720.865190] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.865190] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 720.865190] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 720.865190] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 720.865190] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 720.865190] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 720.865190] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.865190] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 720.865190] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.865190] env[62692]: ERROR nova.compute.manager raise self.value [ 720.865190] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 720.865190] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 720.865190] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.865190] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 720.865881] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.865881] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 720.865881] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e8165560-d358-4cf0-815e-2e3555f92c1f, please check neutron logs for more information. [ 720.865881] env[62692]: ERROR nova.compute.manager [ 720.865881] env[62692]: Traceback (most recent call last): [ 720.865881] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 720.865881] env[62692]: listener.cb(fileno) [ 720.865881] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.865881] env[62692]: result = function(*args, **kwargs) [ 720.865881] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 720.865881] env[62692]: return func(*args, **kwargs) [ 720.865881] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 720.865881] env[62692]: raise e [ 720.865881] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.865881] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 720.865881] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 720.865881] env[62692]: created_port_ids = self._update_ports_for_instance( [ 720.865881] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 720.865881] env[62692]: with excutils.save_and_reraise_exception(): [ 720.865881] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.865881] env[62692]: self.force_reraise() [ 720.865881] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.865881] env[62692]: raise self.value [ 720.865881] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 720.865881] env[62692]: updated_port = self._update_port( [ 720.865881] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.865881] env[62692]: _ensure_no_port_binding_failure(port) [ 720.865881] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.865881] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 720.866919] env[62692]: nova.exception.PortBindingFailed: Binding failed for port e8165560-d358-4cf0-815e-2e3555f92c1f, please check neutron logs for more information. [ 720.866919] env[62692]: Removing descriptor: 17 [ 720.867104] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcfb0cba-897c-4532-a891-28d9a4234b42 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.883642] env[62692]: ERROR nova.compute.manager [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e8165560-d358-4cf0-815e-2e3555f92c1f, please check neutron logs for more information. [ 720.883642] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] Traceback (most recent call last): [ 720.883642] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 720.883642] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] yield resources [ 720.883642] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 720.883642] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] self.driver.spawn(context, instance, image_meta, [ 720.883642] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 720.883642] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] self._vmops.spawn(context, instance, image_meta, injected_files, [ 720.883642] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 720.883642] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] vm_ref = self.build_virtual_machine(instance, [ 720.883642] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 720.884377] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] vif_infos = vmwarevif.get_vif_info(self._session, [ 720.884377] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 720.884377] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] for vif in network_info: [ 720.884377] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 720.884377] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] return self._sync_wrapper(fn, *args, **kwargs) [ 720.884377] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 720.884377] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] self.wait() [ 720.884377] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 720.884377] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] self[:] = self._gt.wait() [ 720.884377] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 720.884377] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] return self._exit_event.wait() [ 720.884377] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 720.884377] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] current.throw(*self._exc) [ 720.884832] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.884832] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] result = function(*args, **kwargs) [ 720.884832] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 720.884832] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] return func(*args, **kwargs) [ 720.884832] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 720.884832] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] raise e [ 720.884832] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.884832] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] nwinfo = self.network_api.allocate_for_instance( [ 720.884832] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 720.884832] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] created_port_ids = self._update_ports_for_instance( [ 720.884832] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 720.884832] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] with excutils.save_and_reraise_exception(): [ 720.884832] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.885223] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] self.force_reraise() [ 720.885223] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.885223] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] raise self.value [ 720.885223] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 720.885223] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] updated_port = self._update_port( [ 720.885223] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.885223] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] _ensure_no_port_binding_failure(port) [ 720.885223] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.885223] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] raise exception.PortBindingFailed(port_id=port['id']) [ 720.885223] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] nova.exception.PortBindingFailed: Binding failed for port e8165560-d358-4cf0-815e-2e3555f92c1f, please check neutron logs for more information. [ 720.885223] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] [ 720.885223] env[62692]: INFO nova.compute.manager [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Terminating instance [ 720.886833] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "refresh_cache-dbd51115-6b41-4019-955d-16a52b534e84" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.101616] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.170967] env[62692]: DEBUG nova.network.neutron [req-59f0f69b-d2d6-4c2d-ad25-09abd30a9a59 req-5f69a6d0-ac68-4136-bba4-f0757cd337f4 service nova] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 721.290923] env[62692]: DEBUG nova.network.neutron [req-59f0f69b-d2d6-4c2d-ad25-09abd30a9a59 req-5f69a6d0-ac68-4136-bba4-f0757cd337f4 service nova] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.345983] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.548s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.346175] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 721.348661] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.693s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.350664] env[62692]: INFO nova.compute.claims [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 721.794007] env[62692]: DEBUG oslo_concurrency.lockutils [req-59f0f69b-d2d6-4c2d-ad25-09abd30a9a59 req-5f69a6d0-ac68-4136-bba4-f0757cd337f4 service nova] Releasing lock "refresh_cache-dbd51115-6b41-4019-955d-16a52b534e84" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.794462] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquired lock "refresh_cache-dbd51115-6b41-4019-955d-16a52b534e84" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.794662] env[62692]: DEBUG nova.network.neutron [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 721.857422] env[62692]: DEBUG nova.compute.utils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 721.858933] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 721.859170] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 721.914998] env[62692]: DEBUG nova.policy [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3ed48d2cdec44fd5ab3175aac2e99747', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eb26b86b1fec4e6fb239b5326cdeae9d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 721.964289] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "8b75e781-930e-4885-81d2-8a8929d6c39b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.964524] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "8b75e781-930e-4885-81d2-8a8929d6c39b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.211868] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Successfully created port: d9b44847-736e-4744-9caa-0c5572ecf152 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 722.316063] env[62692]: DEBUG nova.network.neutron [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 722.363098] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 722.463495] env[62692]: DEBUG nova.network.neutron [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.680144] env[62692]: DEBUG nova.compute.manager [req-7cdccda6-962c-4191-b9b0-a10336d73f5f req-6e8c70ab-d2c2-45fc-bbfc-9d24484789e5 service nova] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Received event network-vif-deleted-e8165560-d358-4cf0-815e-2e3555f92c1f {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 722.799384] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35f9b9c3-abf6-49db-a66b-70692104b79e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.807346] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7a9a4bc-2bf9-437d-8f7c-c132f6c17a62 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.845363] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab809cf-75ba-478d-ba71-943a524a0453 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.852684] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf47bc00-c840-4b6d-8d0f-b1848ac522bd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.865578] env[62692]: DEBUG nova.compute.provider_tree [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.965788] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Releasing lock "refresh_cache-dbd51115-6b41-4019-955d-16a52b534e84" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.966242] env[62692]: DEBUG nova.compute.manager [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 722.966435] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 722.966718] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e668869e-f460-4040-8e4b-fa5d2271ef5c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.975500] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80f18398-abf1-4d93-87ff-a9f2a5c60d47 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.996756] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dbd51115-6b41-4019-955d-16a52b534e84 could not be found. [ 722.996967] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 722.997238] env[62692]: INFO nova.compute.manager [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Took 0.03 seconds to destroy the instance on the hypervisor. [ 722.997486] env[62692]: DEBUG oslo.service.loopingcall [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 722.997692] env[62692]: DEBUG nova.compute.manager [-] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 722.997785] env[62692]: DEBUG nova.network.neutron [-] [instance: dbd51115-6b41-4019-955d-16a52b534e84] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 723.018059] env[62692]: DEBUG nova.network.neutron [-] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 723.265384] env[62692]: ERROR nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d9b44847-736e-4744-9caa-0c5572ecf152, please check neutron logs for more information. [ 723.265384] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 723.265384] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.265384] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 723.265384] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 723.265384] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 723.265384] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 723.265384] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 723.265384] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.265384] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 723.265384] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.265384] env[62692]: ERROR nova.compute.manager raise self.value [ 723.265384] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 723.265384] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 723.265384] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.265384] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 723.265988] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.265988] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 723.265988] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d9b44847-736e-4744-9caa-0c5572ecf152, please check neutron logs for more information. [ 723.265988] env[62692]: ERROR nova.compute.manager [ 723.265988] env[62692]: Traceback (most recent call last): [ 723.265988] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 723.265988] env[62692]: listener.cb(fileno) [ 723.265988] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 723.265988] env[62692]: result = function(*args, **kwargs) [ 723.265988] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 723.265988] env[62692]: return func(*args, **kwargs) [ 723.265988] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 723.265988] env[62692]: raise e [ 723.266879] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.266879] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 723.266944] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 723.266944] env[62692]: created_port_ids = self._update_ports_for_instance( [ 723.266944] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 723.266944] env[62692]: with excutils.save_and_reraise_exception(): [ 723.266944] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.266944] env[62692]: self.force_reraise() [ 723.266944] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.266944] env[62692]: raise self.value [ 723.266944] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 723.266944] env[62692]: updated_port = self._update_port( [ 723.266944] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.266944] env[62692]: _ensure_no_port_binding_failure(port) [ 723.266944] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.266944] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 723.267404] env[62692]: nova.exception.PortBindingFailed: Binding failed for port d9b44847-736e-4744-9caa-0c5572ecf152, please check neutron logs for more information. [ 723.267404] env[62692]: Removing descriptor: 17 [ 723.371780] env[62692]: DEBUG nova.scheduler.client.report [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 723.375908] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 723.399871] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 723.400148] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 723.400333] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 723.400519] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 723.400665] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 723.400810] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 723.401023] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 723.401190] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 723.401356] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 723.401517] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 723.401685] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 723.402566] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f57ef80d-ac54-40df-9d14-b5a81addc8c9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.410881] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42225c3f-9af6-4405-8eac-2531907a7912 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.425652] env[62692]: ERROR nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d9b44847-736e-4744-9caa-0c5572ecf152, please check neutron logs for more information. [ 723.425652] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Traceback (most recent call last): [ 723.425652] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 723.425652] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] yield resources [ 723.425652] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 723.425652] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] self.driver.spawn(context, instance, image_meta, [ 723.425652] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 723.425652] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 723.425652] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 723.425652] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] vm_ref = self.build_virtual_machine(instance, [ 723.425652] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 723.426190] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] vif_infos = vmwarevif.get_vif_info(self._session, [ 723.426190] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 723.426190] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] for vif in network_info: [ 723.426190] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 723.426190] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] return self._sync_wrapper(fn, *args, **kwargs) [ 723.426190] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 723.426190] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] self.wait() [ 723.426190] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 723.426190] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] self[:] = self._gt.wait() [ 723.426190] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 723.426190] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] return self._exit_event.wait() [ 723.426190] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 723.426190] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] current.throw(*self._exc) [ 723.426585] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 723.426585] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] result = function(*args, **kwargs) [ 723.426585] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 723.426585] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] return func(*args, **kwargs) [ 723.426585] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 723.426585] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] raise e [ 723.426585] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.426585] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] nwinfo = self.network_api.allocate_for_instance( [ 723.426585] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 723.426585] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] created_port_ids = self._update_ports_for_instance( [ 723.426585] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 723.426585] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] with excutils.save_and_reraise_exception(): [ 723.426585] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.426995] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] self.force_reraise() [ 723.426995] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.426995] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] raise self.value [ 723.426995] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 723.426995] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] updated_port = self._update_port( [ 723.426995] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.426995] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] _ensure_no_port_binding_failure(port) [ 723.426995] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.426995] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] raise exception.PortBindingFailed(port_id=port['id']) [ 723.426995] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] nova.exception.PortBindingFailed: Binding failed for port d9b44847-736e-4744-9caa-0c5572ecf152, please check neutron logs for more information. [ 723.426995] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] [ 723.426995] env[62692]: INFO nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Terminating instance [ 723.428039] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Acquiring lock "refresh_cache-49d69437-7046-4cb2-914c-8544dca70d3f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.428039] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Acquired lock "refresh_cache-49d69437-7046-4cb2-914c-8544dca70d3f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.428162] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 723.520468] env[62692]: DEBUG nova.network.neutron [-] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.879098] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.530s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.879602] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 723.882191] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.129s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.883654] env[62692]: INFO nova.compute.claims [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 723.946035] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 724.023390] env[62692]: INFO nova.compute.manager [-] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Took 1.03 seconds to deallocate network for instance. [ 724.025772] env[62692]: DEBUG nova.compute.claims [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 724.025861] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.030391] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.390512] env[62692]: DEBUG nova.compute.utils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 724.392169] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 724.392228] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 724.431626] env[62692]: DEBUG nova.policy [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3ed48d2cdec44fd5ab3175aac2e99747', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eb26b86b1fec4e6fb239b5326cdeae9d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 724.533023] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Releasing lock "refresh_cache-49d69437-7046-4cb2-914c-8544dca70d3f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.533584] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 724.533898] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 724.534542] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2010db8c-7e03-4d90-930b-0401fe00a5a2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.543516] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fd232b2-4c0d-4b6f-965b-0d52cb86f172 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.565376] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 49d69437-7046-4cb2-914c-8544dca70d3f could not be found. [ 724.565746] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 724.565814] env[62692]: INFO nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Took 0.03 seconds to destroy the instance on the hypervisor. [ 724.566739] env[62692]: DEBUG oslo.service.loopingcall [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 724.567011] env[62692]: DEBUG nova.compute.manager [-] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 724.567122] env[62692]: DEBUG nova.network.neutron [-] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 724.590436] env[62692]: DEBUG nova.network.neutron [-] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 724.723101] env[62692]: DEBUG nova.compute.manager [req-e1e41849-5889-463a-8396-6b2a03c641f3 req-4b5dc447-c46e-49fb-aa9c-a4735f2ee234 service nova] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Received event network-changed-d9b44847-736e-4744-9caa-0c5572ecf152 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 724.723101] env[62692]: DEBUG nova.compute.manager [req-e1e41849-5889-463a-8396-6b2a03c641f3 req-4b5dc447-c46e-49fb-aa9c-a4735f2ee234 service nova] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Refreshing instance network info cache due to event network-changed-d9b44847-736e-4744-9caa-0c5572ecf152. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 724.723101] env[62692]: DEBUG oslo_concurrency.lockutils [req-e1e41849-5889-463a-8396-6b2a03c641f3 req-4b5dc447-c46e-49fb-aa9c-a4735f2ee234 service nova] Acquiring lock "refresh_cache-49d69437-7046-4cb2-914c-8544dca70d3f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.723101] env[62692]: DEBUG oslo_concurrency.lockutils [req-e1e41849-5889-463a-8396-6b2a03c641f3 req-4b5dc447-c46e-49fb-aa9c-a4735f2ee234 service nova] Acquired lock "refresh_cache-49d69437-7046-4cb2-914c-8544dca70d3f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.723101] env[62692]: DEBUG nova.network.neutron [req-e1e41849-5889-463a-8396-6b2a03c641f3 req-4b5dc447-c46e-49fb-aa9c-a4735f2ee234 service nova] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Refreshing network info cache for port d9b44847-736e-4744-9caa-0c5572ecf152 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 724.743306] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Successfully created port: 3259562d-0513-42ef-8c6a-6e9c0e41f58c {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 724.895396] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 725.096746] env[62692]: DEBUG nova.network.neutron [-] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.247564] env[62692]: DEBUG nova.network.neutron [req-e1e41849-5889-463a-8396-6b2a03c641f3 req-4b5dc447-c46e-49fb-aa9c-a4735f2ee234 service nova] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 725.260967] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e90fc54-4b3d-4e87-821d-83a9a44ee975 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.268463] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ca4038b-08ed-4d59-8da0-501bc0b20f3d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.299422] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97db0680-122a-4343-b887-e4df282c9c2a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.306777] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-790a877d-9eca-47aa-9dd4-ad4702655a43 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.320122] env[62692]: DEBUG nova.compute.provider_tree [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.379947] env[62692]: DEBUG nova.network.neutron [req-e1e41849-5889-463a-8396-6b2a03c641f3 req-4b5dc447-c46e-49fb-aa9c-a4735f2ee234 service nova] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.599660] env[62692]: INFO nova.compute.manager [-] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Took 1.03 seconds to deallocate network for instance. [ 725.602198] env[62692]: DEBUG nova.compute.claims [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 725.602393] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.736789] env[62692]: ERROR nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3259562d-0513-42ef-8c6a-6e9c0e41f58c, please check neutron logs for more information. [ 725.736789] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 725.736789] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.736789] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 725.736789] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 725.736789] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 725.736789] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 725.736789] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 725.736789] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.736789] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 725.736789] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.736789] env[62692]: ERROR nova.compute.manager raise self.value [ 725.736789] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 725.736789] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 725.736789] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.736789] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 725.738624] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.738624] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 725.738624] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3259562d-0513-42ef-8c6a-6e9c0e41f58c, please check neutron logs for more information. [ 725.738624] env[62692]: ERROR nova.compute.manager [ 725.738624] env[62692]: Traceback (most recent call last): [ 725.738624] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 725.738624] env[62692]: listener.cb(fileno) [ 725.738624] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 725.738624] env[62692]: result = function(*args, **kwargs) [ 725.738624] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 725.738624] env[62692]: return func(*args, **kwargs) [ 725.738624] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 725.738624] env[62692]: raise e [ 725.738624] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.738624] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 725.738624] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 725.738624] env[62692]: created_port_ids = self._update_ports_for_instance( [ 725.738624] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 725.738624] env[62692]: with excutils.save_and_reraise_exception(): [ 725.738624] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.738624] env[62692]: self.force_reraise() [ 725.738624] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.738624] env[62692]: raise self.value [ 725.738624] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 725.738624] env[62692]: updated_port = self._update_port( [ 725.738624] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.738624] env[62692]: _ensure_no_port_binding_failure(port) [ 725.738624] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.738624] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 725.739877] env[62692]: nova.exception.PortBindingFailed: Binding failed for port 3259562d-0513-42ef-8c6a-6e9c0e41f58c, please check neutron logs for more information. [ 725.739877] env[62692]: Removing descriptor: 17 [ 725.826144] env[62692]: DEBUG nova.scheduler.client.report [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 725.882721] env[62692]: DEBUG oslo_concurrency.lockutils [req-e1e41849-5889-463a-8396-6b2a03c641f3 req-4b5dc447-c46e-49fb-aa9c-a4735f2ee234 service nova] Releasing lock "refresh_cache-49d69437-7046-4cb2-914c-8544dca70d3f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.882998] env[62692]: DEBUG nova.compute.manager [req-e1e41849-5889-463a-8396-6b2a03c641f3 req-4b5dc447-c46e-49fb-aa9c-a4735f2ee234 service nova] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Received event network-vif-deleted-d9b44847-736e-4744-9caa-0c5572ecf152 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 725.906452] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 725.931667] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 725.931915] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 725.932083] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 725.932269] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 725.932415] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 725.932560] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 725.932763] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 725.932917] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 725.933094] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 725.933518] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 725.933740] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 725.934602] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bf17a28-6d7b-4600-be2c-e0556d980b80 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.942976] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccadfe09-b396-415d-8851-f4cd9d7700dc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.958369] env[62692]: ERROR nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3259562d-0513-42ef-8c6a-6e9c0e41f58c, please check neutron logs for more information. [ 725.958369] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Traceback (most recent call last): [ 725.958369] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 725.958369] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] yield resources [ 725.958369] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 725.958369] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] self.driver.spawn(context, instance, image_meta, [ 725.958369] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 725.958369] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 725.958369] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 725.958369] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] vm_ref = self.build_virtual_machine(instance, [ 725.958369] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 725.958765] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] vif_infos = vmwarevif.get_vif_info(self._session, [ 725.958765] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 725.958765] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] for vif in network_info: [ 725.958765] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 725.958765] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] return self._sync_wrapper(fn, *args, **kwargs) [ 725.958765] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 725.958765] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] self.wait() [ 725.958765] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 725.958765] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] self[:] = self._gt.wait() [ 725.958765] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 725.958765] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] return self._exit_event.wait() [ 725.958765] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 725.958765] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] current.throw(*self._exc) [ 725.959158] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 725.959158] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] result = function(*args, **kwargs) [ 725.959158] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 725.959158] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] return func(*args, **kwargs) [ 725.959158] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 725.959158] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] raise e [ 725.959158] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.959158] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] nwinfo = self.network_api.allocate_for_instance( [ 725.959158] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 725.959158] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] created_port_ids = self._update_ports_for_instance( [ 725.959158] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 725.959158] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] with excutils.save_and_reraise_exception(): [ 725.959158] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.959527] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] self.force_reraise() [ 725.959527] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.959527] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] raise self.value [ 725.959527] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 725.959527] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] updated_port = self._update_port( [ 725.959527] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.959527] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] _ensure_no_port_binding_failure(port) [ 725.959527] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.959527] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] raise exception.PortBindingFailed(port_id=port['id']) [ 725.959527] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] nova.exception.PortBindingFailed: Binding failed for port 3259562d-0513-42ef-8c6a-6e9c0e41f58c, please check neutron logs for more information. [ 725.959527] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] [ 725.959527] env[62692]: INFO nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Terminating instance [ 725.960855] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Acquiring lock "refresh_cache-db727822-af78-49bc-98d9-9caabc35b3b6" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.961015] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Acquired lock "refresh_cache-db727822-af78-49bc-98d9-9caabc35b3b6" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.961185] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 726.331075] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.449s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.331623] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 726.334513] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.442s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.477900] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 726.565373] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.768260] env[62692]: DEBUG nova.compute.manager [req-b0b22156-f621-4e3d-9080-b2cdb6a72d51 req-f49436c3-f413-4df9-b5ce-f0c23c6c54e4 service nova] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Received event network-changed-3259562d-0513-42ef-8c6a-6e9c0e41f58c {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 726.768499] env[62692]: DEBUG nova.compute.manager [req-b0b22156-f621-4e3d-9080-b2cdb6a72d51 req-f49436c3-f413-4df9-b5ce-f0c23c6c54e4 service nova] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Refreshing instance network info cache due to event network-changed-3259562d-0513-42ef-8c6a-6e9c0e41f58c. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 726.768690] env[62692]: DEBUG oslo_concurrency.lockutils [req-b0b22156-f621-4e3d-9080-b2cdb6a72d51 req-f49436c3-f413-4df9-b5ce-f0c23c6c54e4 service nova] Acquiring lock "refresh_cache-db727822-af78-49bc-98d9-9caabc35b3b6" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.839511] env[62692]: DEBUG nova.compute.utils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 726.843527] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 726.843527] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 726.880719] env[62692]: DEBUG nova.policy [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3ed48d2cdec44fd5ab3175aac2e99747', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eb26b86b1fec4e6fb239b5326cdeae9d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 727.069574] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Releasing lock "refresh_cache-db727822-af78-49bc-98d9-9caabc35b3b6" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.070016] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 727.070218] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 727.070527] env[62692]: DEBUG oslo_concurrency.lockutils [req-b0b22156-f621-4e3d-9080-b2cdb6a72d51 req-f49436c3-f413-4df9-b5ce-f0c23c6c54e4 service nova] Acquired lock "refresh_cache-db727822-af78-49bc-98d9-9caabc35b3b6" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.070697] env[62692]: DEBUG nova.network.neutron [req-b0b22156-f621-4e3d-9080-b2cdb6a72d51 req-f49436c3-f413-4df9-b5ce-f0c23c6c54e4 service nova] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Refreshing network info cache for port 3259562d-0513-42ef-8c6a-6e9c0e41f58c {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 727.071737] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d12d1451-9f3f-40be-ada6-e793ba6ca6c2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.086189] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-382e7c28-c2e7-4ef5-8774-53dc8a68e63d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.114109] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance db727822-af78-49bc-98d9-9caabc35b3b6 could not be found. [ 727.114350] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 727.114531] env[62692]: INFO nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 727.114789] env[62692]: DEBUG oslo.service.loopingcall [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 727.115050] env[62692]: DEBUG nova.compute.manager [-] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 727.115164] env[62692]: DEBUG nova.network.neutron [-] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 727.130803] env[62692]: DEBUG nova.network.neutron [-] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 727.191223] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Successfully created port: 8fcbd7c0-2426-4395-96f4-617b331079ab {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 727.248278] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e57aa96-172c-4b2b-ae6c-5a000845dfda {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.255855] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17b71585-36b1-4a80-b8fa-413cc6e0191b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.292416] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0944dfb2-acca-4639-8b39-067cd9c0f56c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.300573] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abd51453-31c9-416d-951e-60f86996a0b4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.314031] env[62692]: DEBUG nova.compute.provider_tree [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.344592] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 727.593624] env[62692]: DEBUG nova.network.neutron [req-b0b22156-f621-4e3d-9080-b2cdb6a72d51 req-f49436c3-f413-4df9-b5ce-f0c23c6c54e4 service nova] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 727.633896] env[62692]: DEBUG nova.network.neutron [-] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.709145] env[62692]: DEBUG nova.network.neutron [req-b0b22156-f621-4e3d-9080-b2cdb6a72d51 req-f49436c3-f413-4df9-b5ce-f0c23c6c54e4 service nova] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.817827] env[62692]: DEBUG nova.scheduler.client.report [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 728.137565] env[62692]: INFO nova.compute.manager [-] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Took 1.02 seconds to deallocate network for instance. [ 728.139974] env[62692]: DEBUG nova.compute.claims [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 728.140165] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.212414] env[62692]: DEBUG oslo_concurrency.lockutils [req-b0b22156-f621-4e3d-9080-b2cdb6a72d51 req-f49436c3-f413-4df9-b5ce-f0c23c6c54e4 service nova] Releasing lock "refresh_cache-db727822-af78-49bc-98d9-9caabc35b3b6" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.212414] env[62692]: DEBUG nova.compute.manager [req-b0b22156-f621-4e3d-9080-b2cdb6a72d51 req-f49436c3-f413-4df9-b5ce-f0c23c6c54e4 service nova] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Received event network-vif-deleted-3259562d-0513-42ef-8c6a-6e9c0e41f58c {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 728.322571] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.988s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.323227] env[62692]: ERROR nova.compute.manager [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port eed49ef6-5628-4bd8-a95e-05d1f3c803cc, please check neutron logs for more information. [ 728.323227] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Traceback (most recent call last): [ 728.323227] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 728.323227] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] self.driver.spawn(context, instance, image_meta, [ 728.323227] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 728.323227] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 728.323227] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 728.323227] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] vm_ref = self.build_virtual_machine(instance, [ 728.323227] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 728.323227] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] vif_infos = vmwarevif.get_vif_info(self._session, [ 728.323227] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 728.323614] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] for vif in network_info: [ 728.323614] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 728.323614] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] return self._sync_wrapper(fn, *args, **kwargs) [ 728.323614] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 728.323614] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] self.wait() [ 728.323614] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 728.323614] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] self[:] = self._gt.wait() [ 728.323614] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 728.323614] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] return self._exit_event.wait() [ 728.323614] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 728.323614] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] result = hub.switch() [ 728.323614] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 728.323614] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] return self.greenlet.switch() [ 728.323926] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.323926] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] result = function(*args, **kwargs) [ 728.323926] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 728.323926] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] return func(*args, **kwargs) [ 728.323926] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 728.323926] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] raise e [ 728.323926] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.323926] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] nwinfo = self.network_api.allocate_for_instance( [ 728.323926] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 728.323926] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] created_port_ids = self._update_ports_for_instance( [ 728.323926] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 728.323926] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] with excutils.save_and_reraise_exception(): [ 728.323926] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.324281] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] self.force_reraise() [ 728.324281] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.324281] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] raise self.value [ 728.324281] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 728.324281] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] updated_port = self._update_port( [ 728.324281] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.324281] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] _ensure_no_port_binding_failure(port) [ 728.324281] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.324281] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] raise exception.PortBindingFailed(port_id=port['id']) [ 728.324281] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] nova.exception.PortBindingFailed: Binding failed for port eed49ef6-5628-4bd8-a95e-05d1f3c803cc, please check neutron logs for more information. [ 728.324281] env[62692]: ERROR nova.compute.manager [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] [ 728.324586] env[62692]: DEBUG nova.compute.utils [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Binding failed for port eed49ef6-5628-4bd8-a95e-05d1f3c803cc, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 728.325275] env[62692]: DEBUG oslo_concurrency.lockutils [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.133s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.326791] env[62692]: INFO nova.compute.claims [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 728.329447] env[62692]: DEBUG nova.compute.manager [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Build of instance 2feffd08-f011-4117-9f8d-ac6730f893f5 was re-scheduled: Binding failed for port eed49ef6-5628-4bd8-a95e-05d1f3c803cc, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 728.330314] env[62692]: DEBUG nova.compute.manager [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 728.330314] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Acquiring lock "refresh_cache-2feffd08-f011-4117-9f8d-ac6730f893f5" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.330314] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Acquired lock "refresh_cache-2feffd08-f011-4117-9f8d-ac6730f893f5" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.330466] env[62692]: DEBUG nova.network.neutron [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 728.352533] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 728.358252] env[62692]: ERROR nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8fcbd7c0-2426-4395-96f4-617b331079ab, please check neutron logs for more information. [ 728.358252] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 728.358252] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.358252] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 728.358252] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 728.358252] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 728.358252] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 728.358252] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 728.358252] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.358252] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 728.358252] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.358252] env[62692]: ERROR nova.compute.manager raise self.value [ 728.358252] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 728.358252] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 728.358252] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.358252] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 728.358665] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.358665] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 728.358665] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8fcbd7c0-2426-4395-96f4-617b331079ab, please check neutron logs for more information. [ 728.358665] env[62692]: ERROR nova.compute.manager [ 728.358665] env[62692]: Traceback (most recent call last): [ 728.358665] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 728.358665] env[62692]: listener.cb(fileno) [ 728.358665] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.358665] env[62692]: result = function(*args, **kwargs) [ 728.358665] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 728.358665] env[62692]: return func(*args, **kwargs) [ 728.358665] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 728.358665] env[62692]: raise e [ 728.358665] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.358665] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 728.358665] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 728.358665] env[62692]: created_port_ids = self._update_ports_for_instance( [ 728.358665] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 728.358665] env[62692]: with excutils.save_and_reraise_exception(): [ 728.358665] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.358665] env[62692]: self.force_reraise() [ 728.358665] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.358665] env[62692]: raise self.value [ 728.358665] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 728.358665] env[62692]: updated_port = self._update_port( [ 728.358665] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.358665] env[62692]: _ensure_no_port_binding_failure(port) [ 728.358665] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.358665] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 728.359479] env[62692]: nova.exception.PortBindingFailed: Binding failed for port 8fcbd7c0-2426-4395-96f4-617b331079ab, please check neutron logs for more information. [ 728.359479] env[62692]: Removing descriptor: 17 [ 728.376777] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 728.377028] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 728.377194] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 728.377380] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 728.377528] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 728.377676] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 728.377881] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 728.378087] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 728.378287] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 728.378454] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 728.378627] env[62692]: DEBUG nova.virt.hardware [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 728.379504] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32036436-ab56-4d1e-be19-4a5bf7475364 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.389019] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20f1f8ee-8d61-4bd2-b73f-1893925c2ff9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.400647] env[62692]: ERROR nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8fcbd7c0-2426-4395-96f4-617b331079ab, please check neutron logs for more information. [ 728.400647] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Traceback (most recent call last): [ 728.400647] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 728.400647] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] yield resources [ 728.400647] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 728.400647] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] self.driver.spawn(context, instance, image_meta, [ 728.400647] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 728.400647] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] self._vmops.spawn(context, instance, image_meta, injected_files, [ 728.400647] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 728.400647] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] vm_ref = self.build_virtual_machine(instance, [ 728.400647] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 728.401011] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] vif_infos = vmwarevif.get_vif_info(self._session, [ 728.401011] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 728.401011] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] for vif in network_info: [ 728.401011] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 728.401011] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] return self._sync_wrapper(fn, *args, **kwargs) [ 728.401011] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 728.401011] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] self.wait() [ 728.401011] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 728.401011] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] self[:] = self._gt.wait() [ 728.401011] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 728.401011] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] return self._exit_event.wait() [ 728.401011] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 728.401011] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] current.throw(*self._exc) [ 728.401354] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.401354] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] result = function(*args, **kwargs) [ 728.401354] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 728.401354] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] return func(*args, **kwargs) [ 728.401354] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 728.401354] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] raise e [ 728.401354] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.401354] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] nwinfo = self.network_api.allocate_for_instance( [ 728.401354] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 728.401354] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] created_port_ids = self._update_ports_for_instance( [ 728.401354] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 728.401354] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] with excutils.save_and_reraise_exception(): [ 728.401354] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.401688] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] self.force_reraise() [ 728.401688] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.401688] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] raise self.value [ 728.401688] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 728.401688] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] updated_port = self._update_port( [ 728.401688] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.401688] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] _ensure_no_port_binding_failure(port) [ 728.401688] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.401688] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] raise exception.PortBindingFailed(port_id=port['id']) [ 728.401688] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] nova.exception.PortBindingFailed: Binding failed for port 8fcbd7c0-2426-4395-96f4-617b331079ab, please check neutron logs for more information. [ 728.401688] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] [ 728.401688] env[62692]: INFO nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Terminating instance [ 728.402809] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Acquiring lock "refresh_cache-0266b67a-7c53-471d-a167-ac7362c4b9be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.402965] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Acquired lock "refresh_cache-0266b67a-7c53-471d-a167-ac7362c4b9be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.403157] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 728.793083] env[62692]: DEBUG nova.compute.manager [req-d31494b8-c82a-4cb2-81fa-5d251db4de0e req-f1f3746a-95f1-4b9b-90cb-436881318c87 service nova] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Received event network-changed-8fcbd7c0-2426-4395-96f4-617b331079ab {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 728.793325] env[62692]: DEBUG nova.compute.manager [req-d31494b8-c82a-4cb2-81fa-5d251db4de0e req-f1f3746a-95f1-4b9b-90cb-436881318c87 service nova] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Refreshing instance network info cache due to event network-changed-8fcbd7c0-2426-4395-96f4-617b331079ab. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 728.793545] env[62692]: DEBUG oslo_concurrency.lockutils [req-d31494b8-c82a-4cb2-81fa-5d251db4de0e req-f1f3746a-95f1-4b9b-90cb-436881318c87 service nova] Acquiring lock "refresh_cache-0266b67a-7c53-471d-a167-ac7362c4b9be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.858525] env[62692]: DEBUG nova.network.neutron [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 728.935854] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 728.975995] env[62692]: DEBUG nova.network.neutron [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.034265] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.479418] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Releasing lock "refresh_cache-2feffd08-f011-4117-9f8d-ac6730f893f5" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.479651] env[62692]: DEBUG nova.compute.manager [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 729.479851] env[62692]: DEBUG nova.compute.manager [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 729.480027] env[62692]: DEBUG nova.network.neutron [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 729.494284] env[62692]: DEBUG nova.network.neutron [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 729.537210] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Releasing lock "refresh_cache-0266b67a-7c53-471d-a167-ac7362c4b9be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.537627] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 729.537820] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 729.538863] env[62692]: DEBUG oslo_concurrency.lockutils [req-d31494b8-c82a-4cb2-81fa-5d251db4de0e req-f1f3746a-95f1-4b9b-90cb-436881318c87 service nova] Acquired lock "refresh_cache-0266b67a-7c53-471d-a167-ac7362c4b9be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.538863] env[62692]: DEBUG nova.network.neutron [req-d31494b8-c82a-4cb2-81fa-5d251db4de0e req-f1f3746a-95f1-4b9b-90cb-436881318c87 service nova] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Refreshing network info cache for port 8fcbd7c0-2426-4395-96f4-617b331079ab {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 729.539613] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-06e4a773-a05b-4051-8036-47a22f5aa0a5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.548935] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a68a2ec4-6c79-4da2-94cb-5ddbc6cc4630 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.573362] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0266b67a-7c53-471d-a167-ac7362c4b9be could not be found. [ 729.573542] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 729.573723] env[62692]: INFO nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Took 0.04 seconds to destroy the instance on the hypervisor. [ 729.573960] env[62692]: DEBUG oslo.service.loopingcall [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 729.576301] env[62692]: DEBUG nova.compute.manager [-] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 729.576402] env[62692]: DEBUG nova.network.neutron [-] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 729.591123] env[62692]: DEBUG nova.network.neutron [-] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 729.705772] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-909cae40-f6a2-40f0-b70e-f3953634fadf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.713044] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-decd1d95-cdca-4d3c-8622-1796634cd03a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.743788] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bf335f4-c98b-4db4-970e-405279c4a683 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.750726] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6985bc7-4558-40f4-b1ca-732da891d078 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.763429] env[62692]: DEBUG nova.compute.provider_tree [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 729.996529] env[62692]: DEBUG nova.network.neutron [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.056744] env[62692]: DEBUG nova.network.neutron [req-d31494b8-c82a-4cb2-81fa-5d251db4de0e req-f1f3746a-95f1-4b9b-90cb-436881318c87 service nova] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 730.092998] env[62692]: DEBUG nova.network.neutron [-] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.136705] env[62692]: DEBUG nova.network.neutron [req-d31494b8-c82a-4cb2-81fa-5d251db4de0e req-f1f3746a-95f1-4b9b-90cb-436881318c87 service nova] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.266744] env[62692]: DEBUG nova.scheduler.client.report [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 730.499416] env[62692]: INFO nova.compute.manager [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] [instance: 2feffd08-f011-4117-9f8d-ac6730f893f5] Took 1.02 seconds to deallocate network for instance. [ 730.597569] env[62692]: INFO nova.compute.manager [-] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Took 1.02 seconds to deallocate network for instance. [ 730.600091] env[62692]: DEBUG nova.compute.claims [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 730.600333] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.639579] env[62692]: DEBUG oslo_concurrency.lockutils [req-d31494b8-c82a-4cb2-81fa-5d251db4de0e req-f1f3746a-95f1-4b9b-90cb-436881318c87 service nova] Releasing lock "refresh_cache-0266b67a-7c53-471d-a167-ac7362c4b9be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.639917] env[62692]: DEBUG nova.compute.manager [req-d31494b8-c82a-4cb2-81fa-5d251db4de0e req-f1f3746a-95f1-4b9b-90cb-436881318c87 service nova] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Received event network-vif-deleted-8fcbd7c0-2426-4395-96f4-617b331079ab {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 730.772090] env[62692]: DEBUG oslo_concurrency.lockutils [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.446s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.773026] env[62692]: DEBUG nova.compute.manager [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 730.775008] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.858s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.776710] env[62692]: INFO nova.compute.claims [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 731.282060] env[62692]: DEBUG nova.compute.utils [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 731.283855] env[62692]: DEBUG nova.compute.manager [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 731.284120] env[62692]: DEBUG nova.network.neutron [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 731.352801] env[62692]: DEBUG nova.policy [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '48184f62238d496996b17f5e15c4fd61', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '17235949812d47f4b82fae2a07c4240a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 731.529020] env[62692]: INFO nova.scheduler.client.report [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Deleted allocations for instance 2feffd08-f011-4117-9f8d-ac6730f893f5 [ 731.685751] env[62692]: DEBUG nova.network.neutron [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Successfully created port: ec8028c0-e992-442a-a555-ea6f3e790f35 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 731.791334] env[62692]: DEBUG nova.compute.manager [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 732.036231] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f9c9adef-808b-48c7-b2ca-b52a760704ff tempest-ServerAddressesTestJSON-909307915 tempest-ServerAddressesTestJSON-909307915-project-member] Lock "2feffd08-f011-4117-9f8d-ac6730f893f5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 92.318s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.249018] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7458c870-5aab-4fb6-be9f-ef62b339f584 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.255860] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94d6dfc4-7038-43ea-af2a-8bdcd74400c9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.285173] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fae7bd13-a327-47e2-91e1-cecdd975ad15 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.292301] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf118f10-f256-4627-9c76-a3b6735f35b5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.314847] env[62692]: DEBUG nova.compute.provider_tree [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 732.540145] env[62692]: DEBUG nova.compute.manager [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 732.625363] env[62692]: DEBUG nova.compute.manager [req-7a78425b-25f9-4c10-8e6b-008a9da80c07 req-1ae95e77-a06a-4d93-b8c2-26988239298d service nova] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Received event network-changed-ec8028c0-e992-442a-a555-ea6f3e790f35 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 732.625363] env[62692]: DEBUG nova.compute.manager [req-7a78425b-25f9-4c10-8e6b-008a9da80c07 req-1ae95e77-a06a-4d93-b8c2-26988239298d service nova] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Refreshing instance network info cache due to event network-changed-ec8028c0-e992-442a-a555-ea6f3e790f35. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 732.625363] env[62692]: DEBUG oslo_concurrency.lockutils [req-7a78425b-25f9-4c10-8e6b-008a9da80c07 req-1ae95e77-a06a-4d93-b8c2-26988239298d service nova] Acquiring lock "refresh_cache-a5fa20ab-66b7-4632-8a80-ac031263c042" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.625363] env[62692]: DEBUG oslo_concurrency.lockutils [req-7a78425b-25f9-4c10-8e6b-008a9da80c07 req-1ae95e77-a06a-4d93-b8c2-26988239298d service nova] Acquired lock "refresh_cache-a5fa20ab-66b7-4632-8a80-ac031263c042" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.625363] env[62692]: DEBUG nova.network.neutron [req-7a78425b-25f9-4c10-8e6b-008a9da80c07 req-1ae95e77-a06a-4d93-b8c2-26988239298d service nova] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Refreshing network info cache for port ec8028c0-e992-442a-a555-ea6f3e790f35 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 732.752276] env[62692]: ERROR nova.compute.manager [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ec8028c0-e992-442a-a555-ea6f3e790f35, please check neutron logs for more information. [ 732.752276] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 732.752276] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.752276] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 732.752276] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 732.752276] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 732.752276] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 732.752276] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 732.752276] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.752276] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 732.752276] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.752276] env[62692]: ERROR nova.compute.manager raise self.value [ 732.752276] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 732.752276] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 732.752276] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.752276] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 732.752977] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.752977] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 732.752977] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ec8028c0-e992-442a-a555-ea6f3e790f35, please check neutron logs for more information. [ 732.752977] env[62692]: ERROR nova.compute.manager [ 732.752977] env[62692]: Traceback (most recent call last): [ 732.752977] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 732.752977] env[62692]: listener.cb(fileno) [ 732.752977] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.752977] env[62692]: result = function(*args, **kwargs) [ 732.752977] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 732.752977] env[62692]: return func(*args, **kwargs) [ 732.752977] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.752977] env[62692]: raise e [ 732.752977] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.752977] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 732.752977] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 732.752977] env[62692]: created_port_ids = self._update_ports_for_instance( [ 732.752977] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 732.752977] env[62692]: with excutils.save_and_reraise_exception(): [ 732.752977] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.752977] env[62692]: self.force_reraise() [ 732.752977] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.752977] env[62692]: raise self.value [ 732.752977] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 732.752977] env[62692]: updated_port = self._update_port( [ 732.752977] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.752977] env[62692]: _ensure_no_port_binding_failure(port) [ 732.752977] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.752977] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 732.754151] env[62692]: nova.exception.PortBindingFailed: Binding failed for port ec8028c0-e992-442a-a555-ea6f3e790f35, please check neutron logs for more information. [ 732.754151] env[62692]: Removing descriptor: 19 [ 732.820627] env[62692]: DEBUG nova.compute.manager [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 732.823240] env[62692]: DEBUG nova.scheduler.client.report [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 732.856702] env[62692]: DEBUG nova.virt.hardware [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 732.856973] env[62692]: DEBUG nova.virt.hardware [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 732.857156] env[62692]: DEBUG nova.virt.hardware [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 732.857343] env[62692]: DEBUG nova.virt.hardware [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 732.857488] env[62692]: DEBUG nova.virt.hardware [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 732.857634] env[62692]: DEBUG nova.virt.hardware [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 732.857852] env[62692]: DEBUG nova.virt.hardware [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 732.858044] env[62692]: DEBUG nova.virt.hardware [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 732.858236] env[62692]: DEBUG nova.virt.hardware [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 732.858401] env[62692]: DEBUG nova.virt.hardware [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 732.858596] env[62692]: DEBUG nova.virt.hardware [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 732.859691] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c798694e-e5fe-43c4-8c96-2909de8e653d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.868071] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00daa6cc-452e-4216-a68c-7860989ee6f4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.881512] env[62692]: ERROR nova.compute.manager [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ec8028c0-e992-442a-a555-ea6f3e790f35, please check neutron logs for more information. [ 732.881512] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Traceback (most recent call last): [ 732.881512] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 732.881512] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] yield resources [ 732.881512] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 732.881512] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] self.driver.spawn(context, instance, image_meta, [ 732.881512] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 732.881512] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] self._vmops.spawn(context, instance, image_meta, injected_files, [ 732.881512] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 732.881512] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] vm_ref = self.build_virtual_machine(instance, [ 732.881512] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 732.881930] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] vif_infos = vmwarevif.get_vif_info(self._session, [ 732.881930] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 732.881930] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] for vif in network_info: [ 732.881930] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 732.881930] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] return self._sync_wrapper(fn, *args, **kwargs) [ 732.881930] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 732.881930] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] self.wait() [ 732.881930] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 732.881930] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] self[:] = self._gt.wait() [ 732.881930] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 732.881930] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] return self._exit_event.wait() [ 732.881930] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 732.881930] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] current.throw(*self._exc) [ 732.882332] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.882332] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] result = function(*args, **kwargs) [ 732.882332] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 732.882332] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] return func(*args, **kwargs) [ 732.882332] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.882332] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] raise e [ 732.882332] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.882332] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] nwinfo = self.network_api.allocate_for_instance( [ 732.882332] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 732.882332] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] created_port_ids = self._update_ports_for_instance( [ 732.882332] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 732.882332] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] with excutils.save_and_reraise_exception(): [ 732.882332] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.882704] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] self.force_reraise() [ 732.882704] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.882704] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] raise self.value [ 732.882704] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 732.882704] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] updated_port = self._update_port( [ 732.882704] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.882704] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] _ensure_no_port_binding_failure(port) [ 732.882704] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.882704] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] raise exception.PortBindingFailed(port_id=port['id']) [ 732.882704] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] nova.exception.PortBindingFailed: Binding failed for port ec8028c0-e992-442a-a555-ea6f3e790f35, please check neutron logs for more information. [ 732.882704] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] [ 732.882704] env[62692]: INFO nova.compute.manager [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Terminating instance [ 732.883912] env[62692]: DEBUG oslo_concurrency.lockutils [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Acquiring lock "refresh_cache-a5fa20ab-66b7-4632-8a80-ac031263c042" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.063565] env[62692]: DEBUG oslo_concurrency.lockutils [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.143357] env[62692]: DEBUG nova.network.neutron [req-7a78425b-25f9-4c10-8e6b-008a9da80c07 req-1ae95e77-a06a-4d93-b8c2-26988239298d service nova] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 733.253169] env[62692]: DEBUG nova.network.neutron [req-7a78425b-25f9-4c10-8e6b-008a9da80c07 req-1ae95e77-a06a-4d93-b8c2-26988239298d service nova] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.328559] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.553s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.329144] env[62692]: DEBUG nova.compute.manager [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 733.332870] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.166s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.759647] env[62692]: DEBUG oslo_concurrency.lockutils [req-7a78425b-25f9-4c10-8e6b-008a9da80c07 req-1ae95e77-a06a-4d93-b8c2-26988239298d service nova] Releasing lock "refresh_cache-a5fa20ab-66b7-4632-8a80-ac031263c042" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.760092] env[62692]: DEBUG oslo_concurrency.lockutils [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Acquired lock "refresh_cache-a5fa20ab-66b7-4632-8a80-ac031263c042" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.760292] env[62692]: DEBUG nova.network.neutron [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 733.837740] env[62692]: DEBUG nova.compute.utils [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 733.842435] env[62692]: DEBUG nova.compute.manager [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 733.842621] env[62692]: DEBUG nova.network.neutron [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 733.902373] env[62692]: DEBUG nova.policy [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0b68042fb3344097b2b1567e7a032bd1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '25d21356002248e18d587a77c05ebe82', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 734.290455] env[62692]: DEBUG nova.network.neutron [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 734.302351] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f395d28e-4042-40ee-bce0-0086a1fa1b82 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.311205] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74ea2c3b-6f4b-4b06-bfd2-633dd41f67cc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.318424] env[62692]: DEBUG nova.network.neutron [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Successfully created port: 3736cd8e-be04-4c79-985a-54953ea3d16c {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 734.347713] env[62692]: DEBUG nova.compute.manager [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 734.353904] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afa7c3d6-84c4-4de5-a177-cb66898eef3d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.362297] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bc87fa5-f383-41ef-8d96-9a60c87aa53a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.376879] env[62692]: DEBUG nova.compute.provider_tree [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.427727] env[62692]: DEBUG nova.network.neutron [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.650725] env[62692]: DEBUG nova.compute.manager [req-c90feeff-4751-4c47-8f54-f1d11e843746 req-c52febbf-519e-488d-9c4b-fef57bc75a45 service nova] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Received event network-vif-deleted-ec8028c0-e992-442a-a555-ea6f3e790f35 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 734.881232] env[62692]: DEBUG nova.scheduler.client.report [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 734.930512] env[62692]: DEBUG oslo_concurrency.lockutils [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Releasing lock "refresh_cache-a5fa20ab-66b7-4632-8a80-ac031263c042" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.930944] env[62692]: DEBUG nova.compute.manager [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 734.931157] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 734.931519] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aa79c637-6f1d-446b-928b-3692ee830b30 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.941768] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-618f7750-ce52-4516-825c-c9b4a1bc39f8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.966746] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a5fa20ab-66b7-4632-8a80-ac031263c042 could not be found. [ 734.966985] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 734.967182] env[62692]: INFO nova.compute.manager [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Took 0.04 seconds to destroy the instance on the hypervisor. [ 734.967430] env[62692]: DEBUG oslo.service.loopingcall [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 734.967648] env[62692]: DEBUG nova.compute.manager [-] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 734.967755] env[62692]: DEBUG nova.network.neutron [-] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 734.986837] env[62692]: DEBUG nova.network.neutron [-] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 735.356359] env[62692]: ERROR nova.compute.manager [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3736cd8e-be04-4c79-985a-54953ea3d16c, please check neutron logs for more information. [ 735.356359] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 735.356359] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.356359] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 735.356359] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 735.356359] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 735.356359] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 735.356359] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 735.356359] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.356359] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 735.356359] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.356359] env[62692]: ERROR nova.compute.manager raise self.value [ 735.356359] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 735.356359] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 735.356359] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.356359] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 735.356986] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.356986] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 735.356986] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3736cd8e-be04-4c79-985a-54953ea3d16c, please check neutron logs for more information. [ 735.356986] env[62692]: ERROR nova.compute.manager [ 735.356986] env[62692]: Traceback (most recent call last): [ 735.356986] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 735.356986] env[62692]: listener.cb(fileno) [ 735.356986] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 735.356986] env[62692]: result = function(*args, **kwargs) [ 735.356986] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 735.356986] env[62692]: return func(*args, **kwargs) [ 735.356986] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 735.356986] env[62692]: raise e [ 735.356986] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.356986] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 735.356986] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 735.356986] env[62692]: created_port_ids = self._update_ports_for_instance( [ 735.356986] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 735.356986] env[62692]: with excutils.save_and_reraise_exception(): [ 735.356986] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.356986] env[62692]: self.force_reraise() [ 735.356986] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.356986] env[62692]: raise self.value [ 735.356986] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 735.356986] env[62692]: updated_port = self._update_port( [ 735.356986] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.356986] env[62692]: _ensure_no_port_binding_failure(port) [ 735.356986] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.356986] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 735.357777] env[62692]: nova.exception.PortBindingFailed: Binding failed for port 3736cd8e-be04-4c79-985a-54953ea3d16c, please check neutron logs for more information. [ 735.357777] env[62692]: Removing descriptor: 19 [ 735.357777] env[62692]: DEBUG nova.compute.manager [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 735.384486] env[62692]: DEBUG nova.virt.hardware [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 735.384738] env[62692]: DEBUG nova.virt.hardware [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 735.384896] env[62692]: DEBUG nova.virt.hardware [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 735.385090] env[62692]: DEBUG nova.virt.hardware [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 735.385358] env[62692]: DEBUG nova.virt.hardware [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 735.385512] env[62692]: DEBUG nova.virt.hardware [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 735.385716] env[62692]: DEBUG nova.virt.hardware [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 735.385873] env[62692]: DEBUG nova.virt.hardware [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 735.386057] env[62692]: DEBUG nova.virt.hardware [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 735.386261] env[62692]: DEBUG nova.virt.hardware [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 735.386447] env[62692]: DEBUG nova.virt.hardware [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 735.387177] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.054s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.387761] env[62692]: ERROR nova.compute.manager [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c6d0b984-14e0-4bea-b8d5-82c7df1084b3, please check neutron logs for more information. [ 735.387761] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Traceback (most recent call last): [ 735.387761] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 735.387761] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] self.driver.spawn(context, instance, image_meta, [ 735.387761] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 735.387761] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] self._vmops.spawn(context, instance, image_meta, injected_files, [ 735.387761] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 735.387761] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] vm_ref = self.build_virtual_machine(instance, [ 735.387761] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 735.387761] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] vif_infos = vmwarevif.get_vif_info(self._session, [ 735.387761] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 735.388243] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] for vif in network_info: [ 735.388243] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 735.388243] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] return self._sync_wrapper(fn, *args, **kwargs) [ 735.388243] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 735.388243] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] self.wait() [ 735.388243] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 735.388243] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] self[:] = self._gt.wait() [ 735.388243] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 735.388243] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] return self._exit_event.wait() [ 735.388243] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 735.388243] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] result = hub.switch() [ 735.388243] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 735.388243] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] return self.greenlet.switch() [ 735.388564] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 735.388564] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] result = function(*args, **kwargs) [ 735.388564] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 735.388564] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] return func(*args, **kwargs) [ 735.388564] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 735.388564] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] raise e [ 735.388564] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.388564] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] nwinfo = self.network_api.allocate_for_instance( [ 735.388564] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 735.388564] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] created_port_ids = self._update_ports_for_instance( [ 735.388564] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 735.388564] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] with excutils.save_and_reraise_exception(): [ 735.388564] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.388867] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] self.force_reraise() [ 735.388867] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.388867] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] raise self.value [ 735.388867] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 735.388867] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] updated_port = self._update_port( [ 735.388867] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.388867] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] _ensure_no_port_binding_failure(port) [ 735.388867] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.388867] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] raise exception.PortBindingFailed(port_id=port['id']) [ 735.388867] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] nova.exception.PortBindingFailed: Binding failed for port c6d0b984-14e0-4bea-b8d5-82c7df1084b3, please check neutron logs for more information. [ 735.388867] env[62692]: ERROR nova.compute.manager [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] [ 735.389283] env[62692]: DEBUG nova.compute.utils [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Binding failed for port c6d0b984-14e0-4bea-b8d5-82c7df1084b3, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 735.390271] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d34dd2b0-92c9-4a18-927d-c7f943da396d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.393079] env[62692]: DEBUG nova.compute.manager [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Build of instance 9422a9a8-795d-4794-8ba9-9e6f88cdb464 was re-scheduled: Binding failed for port c6d0b984-14e0-4bea-b8d5-82c7df1084b3, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 735.393503] env[62692]: DEBUG nova.compute.manager [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 735.393732] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Acquiring lock "refresh_cache-9422a9a8-795d-4794-8ba9-9e6f88cdb464" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.393878] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Acquired lock "refresh_cache-9422a9a8-795d-4794-8ba9-9e6f88cdb464" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.394048] env[62692]: DEBUG nova.network.neutron [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 735.395016] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.800s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.409950] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ab2b19a-997d-4a41-b886-610f43a19b8c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.428123] env[62692]: ERROR nova.compute.manager [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3736cd8e-be04-4c79-985a-54953ea3d16c, please check neutron logs for more information. [ 735.428123] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Traceback (most recent call last): [ 735.428123] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 735.428123] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] yield resources [ 735.428123] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 735.428123] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] self.driver.spawn(context, instance, image_meta, [ 735.428123] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 735.428123] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 735.428123] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 735.428123] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] vm_ref = self.build_virtual_machine(instance, [ 735.428123] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 735.428617] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] vif_infos = vmwarevif.get_vif_info(self._session, [ 735.428617] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 735.428617] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] for vif in network_info: [ 735.428617] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 735.428617] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] return self._sync_wrapper(fn, *args, **kwargs) [ 735.428617] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 735.428617] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] self.wait() [ 735.428617] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 735.428617] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] self[:] = self._gt.wait() [ 735.428617] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 735.428617] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] return self._exit_event.wait() [ 735.428617] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 735.428617] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] current.throw(*self._exc) [ 735.428985] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 735.428985] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] result = function(*args, **kwargs) [ 735.428985] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 735.428985] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] return func(*args, **kwargs) [ 735.428985] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 735.428985] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] raise e [ 735.428985] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.428985] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] nwinfo = self.network_api.allocate_for_instance( [ 735.428985] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 735.428985] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] created_port_ids = self._update_ports_for_instance( [ 735.428985] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 735.428985] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] with excutils.save_and_reraise_exception(): [ 735.428985] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.429331] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] self.force_reraise() [ 735.429331] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.429331] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] raise self.value [ 735.429331] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 735.429331] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] updated_port = self._update_port( [ 735.429331] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.429331] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] _ensure_no_port_binding_failure(port) [ 735.429331] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.429331] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] raise exception.PortBindingFailed(port_id=port['id']) [ 735.429331] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] nova.exception.PortBindingFailed: Binding failed for port 3736cd8e-be04-4c79-985a-54953ea3d16c, please check neutron logs for more information. [ 735.429331] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] [ 735.429331] env[62692]: INFO nova.compute.manager [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Terminating instance [ 735.431036] env[62692]: DEBUG nova.network.neutron [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 735.432885] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Acquiring lock "refresh_cache-ed43eb23-0f75-4a8b-96a1-142e1abe400f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.433032] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Acquired lock "refresh_cache-ed43eb23-0f75-4a8b-96a1-142e1abe400f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.433174] env[62692]: DEBUG nova.network.neutron [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 735.489437] env[62692]: DEBUG nova.network.neutron [-] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.501105] env[62692]: DEBUG nova.network.neutron [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.960601] env[62692]: DEBUG nova.network.neutron [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 735.992136] env[62692]: INFO nova.compute.manager [-] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Took 1.02 seconds to deallocate network for instance. [ 735.994516] env[62692]: DEBUG nova.compute.claims [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 735.995122] env[62692]: DEBUG oslo_concurrency.lockutils [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.004274] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Releasing lock "refresh_cache-9422a9a8-795d-4794-8ba9-9e6f88cdb464" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.004488] env[62692]: DEBUG nova.compute.manager [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 736.004652] env[62692]: DEBUG nova.compute.manager [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 736.004820] env[62692]: DEBUG nova.network.neutron [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 736.027496] env[62692]: DEBUG nova.network.neutron [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 736.064526] env[62692]: DEBUG nova.network.neutron [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.234171] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfecfb1c-105b-41c8-a8fb-43476f841792 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.241908] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb60638b-88a6-4cb0-bacb-5f788a6c2458 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.272804] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da4e0185-b18c-4864-98a6-e9ade6b31b3e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.280484] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e9bae06-e918-4e2f-85ae-1295386b834d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.294098] env[62692]: DEBUG nova.compute.provider_tree [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 736.530204] env[62692]: DEBUG nova.network.neutron [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.567045] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Releasing lock "refresh_cache-ed43eb23-0f75-4a8b-96a1-142e1abe400f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.567489] env[62692]: DEBUG nova.compute.manager [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 736.567679] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 736.567953] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3fae9e40-9a2e-4aa2-af0d-e3a1fe5a1f26 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.577741] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0f1f671-26e1-4981-b8b9-8b3000b31b9c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.598830] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ed43eb23-0f75-4a8b-96a1-142e1abe400f could not be found. [ 736.599065] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 736.599364] env[62692]: INFO nova.compute.manager [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Took 0.03 seconds to destroy the instance on the hypervisor. [ 736.599629] env[62692]: DEBUG oslo.service.loopingcall [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 736.599855] env[62692]: DEBUG nova.compute.manager [-] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 736.599951] env[62692]: DEBUG nova.network.neutron [-] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 736.615565] env[62692]: DEBUG nova.network.neutron [-] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 736.673247] env[62692]: DEBUG nova.compute.manager [req-17d76662-3703-428e-8c13-cf2160419730 req-b02e3e8a-26f7-4f4c-8ff3-3c76648a2d29 service nova] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Received event network-changed-3736cd8e-be04-4c79-985a-54953ea3d16c {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 736.673484] env[62692]: DEBUG nova.compute.manager [req-17d76662-3703-428e-8c13-cf2160419730 req-b02e3e8a-26f7-4f4c-8ff3-3c76648a2d29 service nova] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Refreshing instance network info cache due to event network-changed-3736cd8e-be04-4c79-985a-54953ea3d16c. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 736.673653] env[62692]: DEBUG oslo_concurrency.lockutils [req-17d76662-3703-428e-8c13-cf2160419730 req-b02e3e8a-26f7-4f4c-8ff3-3c76648a2d29 service nova] Acquiring lock "refresh_cache-ed43eb23-0f75-4a8b-96a1-142e1abe400f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.673797] env[62692]: DEBUG oslo_concurrency.lockutils [req-17d76662-3703-428e-8c13-cf2160419730 req-b02e3e8a-26f7-4f4c-8ff3-3c76648a2d29 service nova] Acquired lock "refresh_cache-ed43eb23-0f75-4a8b-96a1-142e1abe400f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.673958] env[62692]: DEBUG nova.network.neutron [req-17d76662-3703-428e-8c13-cf2160419730 req-b02e3e8a-26f7-4f4c-8ff3-3c76648a2d29 service nova] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Refreshing network info cache for port 3736cd8e-be04-4c79-985a-54953ea3d16c {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 736.796990] env[62692]: DEBUG nova.scheduler.client.report [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 737.032920] env[62692]: INFO nova.compute.manager [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] [instance: 9422a9a8-795d-4794-8ba9-9e6f88cdb464] Took 1.03 seconds to deallocate network for instance. [ 737.118678] env[62692]: DEBUG nova.network.neutron [-] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.189775] env[62692]: DEBUG nova.network.neutron [req-17d76662-3703-428e-8c13-cf2160419730 req-b02e3e8a-26f7-4f4c-8ff3-3c76648a2d29 service nova] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 737.258325] env[62692]: DEBUG nova.network.neutron [req-17d76662-3703-428e-8c13-cf2160419730 req-b02e3e8a-26f7-4f4c-8ff3-3c76648a2d29 service nova] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.301652] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.907s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.302292] env[62692]: ERROR nova.compute.manager [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1c476cc3-b128-4738-a1d2-8d915dfee19b, please check neutron logs for more information. [ 737.302292] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Traceback (most recent call last): [ 737.302292] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 737.302292] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] self.driver.spawn(context, instance, image_meta, [ 737.302292] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 737.302292] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] self._vmops.spawn(context, instance, image_meta, injected_files, [ 737.302292] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 737.302292] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] vm_ref = self.build_virtual_machine(instance, [ 737.302292] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 737.302292] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] vif_infos = vmwarevif.get_vif_info(self._session, [ 737.302292] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 737.302614] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] for vif in network_info: [ 737.302614] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 737.302614] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] return self._sync_wrapper(fn, *args, **kwargs) [ 737.302614] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 737.302614] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] self.wait() [ 737.302614] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 737.302614] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] self[:] = self._gt.wait() [ 737.302614] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 737.302614] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] return self._exit_event.wait() [ 737.302614] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 737.302614] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] result = hub.switch() [ 737.302614] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 737.302614] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] return self.greenlet.switch() [ 737.302937] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 737.302937] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] result = function(*args, **kwargs) [ 737.302937] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 737.302937] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] return func(*args, **kwargs) [ 737.302937] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 737.302937] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] raise e [ 737.302937] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 737.302937] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] nwinfo = self.network_api.allocate_for_instance( [ 737.302937] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 737.302937] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] created_port_ids = self._update_ports_for_instance( [ 737.302937] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 737.302937] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] with excutils.save_and_reraise_exception(): [ 737.302937] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.303398] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] self.force_reraise() [ 737.303398] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.303398] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] raise self.value [ 737.303398] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 737.303398] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] updated_port = self._update_port( [ 737.303398] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.303398] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] _ensure_no_port_binding_failure(port) [ 737.303398] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.303398] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] raise exception.PortBindingFailed(port_id=port['id']) [ 737.303398] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] nova.exception.PortBindingFailed: Binding failed for port 1c476cc3-b128-4738-a1d2-8d915dfee19b, please check neutron logs for more information. [ 737.303398] env[62692]: ERROR nova.compute.manager [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] [ 737.303703] env[62692]: DEBUG nova.compute.utils [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Binding failed for port 1c476cc3-b128-4738-a1d2-8d915dfee19b, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 737.304211] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 17.038s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.304378] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.304527] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62692) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 737.304798] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.203s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.306236] env[62692]: INFO nova.compute.claims [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 737.309167] env[62692]: DEBUG nova.compute.manager [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Build of instance 146e164e-2543-46ab-a7d3-9803d56f61ec was re-scheduled: Binding failed for port 1c476cc3-b128-4738-a1d2-8d915dfee19b, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 737.309683] env[62692]: DEBUG nova.compute.manager [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 737.309907] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquiring lock "refresh_cache-146e164e-2543-46ab-a7d3-9803d56f61ec" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.310245] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquired lock "refresh_cache-146e164e-2543-46ab-a7d3-9803d56f61ec" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.310427] env[62692]: DEBUG nova.network.neutron [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 737.311834] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-970267ba-6bd2-4a97-a8c1-7656d203f71d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.320901] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38d4eb3d-4734-4ed5-ba79-d1f8ae888009 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.335596] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e720f6e-9b50-4446-b022-6401211a86a4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.342182] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e7c9e46-054b-4e49-8148-8b5addd50b99 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.373216] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181513MB free_disk=167GB free_vcpus=48 pci_devices=None {{(pid=62692) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 737.373432] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.621025] env[62692]: INFO nova.compute.manager [-] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Took 1.02 seconds to deallocate network for instance. [ 737.623287] env[62692]: DEBUG nova.compute.claims [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 737.623542] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.760821] env[62692]: DEBUG oslo_concurrency.lockutils [req-17d76662-3703-428e-8c13-cf2160419730 req-b02e3e8a-26f7-4f4c-8ff3-3c76648a2d29 service nova] Releasing lock "refresh_cache-ed43eb23-0f75-4a8b-96a1-142e1abe400f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.761168] env[62692]: DEBUG nova.compute.manager [req-17d76662-3703-428e-8c13-cf2160419730 req-b02e3e8a-26f7-4f4c-8ff3-3c76648a2d29 service nova] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Received event network-vif-deleted-3736cd8e-be04-4c79-985a-54953ea3d16c {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 737.837464] env[62692]: DEBUG nova.network.neutron [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 737.893026] env[62692]: DEBUG nova.network.neutron [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.061366] env[62692]: INFO nova.scheduler.client.report [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Deleted allocations for instance 9422a9a8-795d-4794-8ba9-9e6f88cdb464 [ 738.395821] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Releasing lock "refresh_cache-146e164e-2543-46ab-a7d3-9803d56f61ec" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.396094] env[62692]: DEBUG nova.compute.manager [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 738.396264] env[62692]: DEBUG nova.compute.manager [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 738.396623] env[62692]: DEBUG nova.network.neutron [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 738.411250] env[62692]: DEBUG nova.network.neutron [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 738.568608] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a981eaad-688d-48ed-834c-475bb9d2f508 tempest-DeleteServersAdminTestJSON-1041233798 tempest-DeleteServersAdminTestJSON-1041233798-project-member] Lock "9422a9a8-795d-4794-8ba9-9e6f88cdb464" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.850s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.648992] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edbdc1f3-a556-408f-b6ee-23d3ba2a4d0e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.656488] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3e89b6d-4d63-4b34-b666-d0706fb4e8a1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.688053] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dde8133-fd97-4757-8ecb-cbb23961e25d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.694534] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8ead5c9-be2a-4eab-8d20-71d4d7472782 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.708115] env[62692]: DEBUG nova.compute.provider_tree [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 738.915957] env[62692]: DEBUG nova.network.neutron [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.071163] env[62692]: DEBUG nova.compute.manager [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 739.211294] env[62692]: DEBUG nova.scheduler.client.report [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 739.420631] env[62692]: INFO nova.compute.manager [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 146e164e-2543-46ab-a7d3-9803d56f61ec] Took 1.02 seconds to deallocate network for instance. [ 739.598372] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.717210] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.411s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.717210] env[62692]: DEBUG nova.compute.manager [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 739.719863] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.694s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.224393] env[62692]: DEBUG nova.compute.utils [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 740.229477] env[62692]: DEBUG nova.compute.manager [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 740.229477] env[62692]: DEBUG nova.network.neutron [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 740.326699] env[62692]: DEBUG nova.policy [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '60e97208df394761848248fca8099364', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e21467aeb32e4b3a9644c35377bbf063', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 740.456018] env[62692]: INFO nova.scheduler.client.report [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Deleted allocations for instance 146e164e-2543-46ab-a7d3-9803d56f61ec [ 740.677150] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-443013d0-bbd9-4b29-a094-db922d018e3d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.686281] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1f3eb4f-cb55-42ff-8ef9-4f897934acf6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.719178] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d42d6a76-e5ce-4b4d-94d4-3a1501109fe3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.729021] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cab986fe-9268-4e53-9d29-3c57a15afdba {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.731958] env[62692]: DEBUG nova.compute.manager [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 740.735103] env[62692]: DEBUG nova.network.neutron [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Successfully created port: 42704cff-466f-442d-8911-a6365882ee2d {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 740.748104] env[62692]: DEBUG nova.compute.provider_tree [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 740.960316] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e5b22dfb-1365-4d16-ad81-531da7ce55c6 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Lock "146e164e-2543-46ab-a7d3-9803d56f61ec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 100.662s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.251198] env[62692]: DEBUG nova.scheduler.client.report [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 741.464296] env[62692]: DEBUG nova.compute.manager [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 741.557246] env[62692]: DEBUG nova.compute.manager [req-d4d945b7-ee43-4f3c-94ee-4221bcf2b0fa req-2fb3c713-2c17-48eb-9665-dbe2537242b8 service nova] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Received event network-changed-42704cff-466f-442d-8911-a6365882ee2d {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 741.557494] env[62692]: DEBUG nova.compute.manager [req-d4d945b7-ee43-4f3c-94ee-4221bcf2b0fa req-2fb3c713-2c17-48eb-9665-dbe2537242b8 service nova] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Refreshing instance network info cache due to event network-changed-42704cff-466f-442d-8911-a6365882ee2d. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 741.557856] env[62692]: DEBUG oslo_concurrency.lockutils [req-d4d945b7-ee43-4f3c-94ee-4221bcf2b0fa req-2fb3c713-2c17-48eb-9665-dbe2537242b8 service nova] Acquiring lock "refresh_cache-700d3723-f787-4770-a409-642cd8085e9c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.557950] env[62692]: DEBUG oslo_concurrency.lockutils [req-d4d945b7-ee43-4f3c-94ee-4221bcf2b0fa req-2fb3c713-2c17-48eb-9665-dbe2537242b8 service nova] Acquired lock "refresh_cache-700d3723-f787-4770-a409-642cd8085e9c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.558174] env[62692]: DEBUG nova.network.neutron [req-d4d945b7-ee43-4f3c-94ee-4221bcf2b0fa req-2fb3c713-2c17-48eb-9665-dbe2537242b8 service nova] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Refreshing network info cache for port 42704cff-466f-442d-8911-a6365882ee2d {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 741.745447] env[62692]: DEBUG nova.compute.manager [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 741.756660] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.037s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.757299] env[62692]: ERROR nova.compute.manager [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e8165560-d358-4cf0-815e-2e3555f92c1f, please check neutron logs for more information. [ 741.757299] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] Traceback (most recent call last): [ 741.757299] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 741.757299] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] self.driver.spawn(context, instance, image_meta, [ 741.757299] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 741.757299] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] self._vmops.spawn(context, instance, image_meta, injected_files, [ 741.757299] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 741.757299] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] vm_ref = self.build_virtual_machine(instance, [ 741.757299] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 741.757299] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] vif_infos = vmwarevif.get_vif_info(self._session, [ 741.757299] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 741.757650] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] for vif in network_info: [ 741.757650] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 741.757650] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] return self._sync_wrapper(fn, *args, **kwargs) [ 741.757650] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 741.757650] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] self.wait() [ 741.757650] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 741.757650] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] self[:] = self._gt.wait() [ 741.757650] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 741.757650] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] return self._exit_event.wait() [ 741.757650] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 741.757650] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] current.throw(*self._exc) [ 741.757650] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 741.757650] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] result = function(*args, **kwargs) [ 741.758070] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 741.758070] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] return func(*args, **kwargs) [ 741.758070] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 741.758070] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] raise e [ 741.758070] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 741.758070] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] nwinfo = self.network_api.allocate_for_instance( [ 741.758070] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 741.758070] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] created_port_ids = self._update_ports_for_instance( [ 741.758070] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 741.758070] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] with excutils.save_and_reraise_exception(): [ 741.758070] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 741.758070] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] self.force_reraise() [ 741.758070] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 741.758462] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] raise self.value [ 741.758462] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 741.758462] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] updated_port = self._update_port( [ 741.758462] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 741.758462] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] _ensure_no_port_binding_failure(port) [ 741.758462] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 741.758462] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] raise exception.PortBindingFailed(port_id=port['id']) [ 741.758462] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] nova.exception.PortBindingFailed: Binding failed for port e8165560-d358-4cf0-815e-2e3555f92c1f, please check neutron logs for more information. [ 741.758462] env[62692]: ERROR nova.compute.manager [instance: dbd51115-6b41-4019-955d-16a52b534e84] [ 741.758462] env[62692]: DEBUG nova.compute.utils [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Binding failed for port e8165560-d358-4cf0-815e-2e3555f92c1f, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 741.759396] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.157s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.762289] env[62692]: DEBUG nova.compute.manager [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Build of instance dbd51115-6b41-4019-955d-16a52b534e84 was re-scheduled: Binding failed for port e8165560-d358-4cf0-815e-2e3555f92c1f, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 741.762710] env[62692]: DEBUG nova.compute.manager [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 741.762924] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "refresh_cache-dbd51115-6b41-4019-955d-16a52b534e84" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.764940] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquired lock "refresh_cache-dbd51115-6b41-4019-955d-16a52b534e84" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.764940] env[62692]: DEBUG nova.network.neutron [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 741.778856] env[62692]: ERROR nova.compute.manager [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 42704cff-466f-442d-8911-a6365882ee2d, please check neutron logs for more information. [ 741.778856] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 741.778856] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 741.778856] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 741.778856] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 741.778856] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 741.778856] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 741.778856] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 741.778856] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 741.778856] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 741.778856] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 741.778856] env[62692]: ERROR nova.compute.manager raise self.value [ 741.778856] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 741.778856] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 741.778856] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 741.778856] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 741.779381] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 741.779381] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 741.779381] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 42704cff-466f-442d-8911-a6365882ee2d, please check neutron logs for more information. [ 741.779381] env[62692]: ERROR nova.compute.manager [ 741.779381] env[62692]: Traceback (most recent call last): [ 741.779381] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 741.779381] env[62692]: listener.cb(fileno) [ 741.779381] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 741.779381] env[62692]: result = function(*args, **kwargs) [ 741.779381] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 741.779381] env[62692]: return func(*args, **kwargs) [ 741.779381] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 741.779381] env[62692]: raise e [ 741.779381] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 741.779381] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 741.779381] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 741.779381] env[62692]: created_port_ids = self._update_ports_for_instance( [ 741.779381] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 741.779381] env[62692]: with excutils.save_and_reraise_exception(): [ 741.779381] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 741.779381] env[62692]: self.force_reraise() [ 741.779381] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 741.779381] env[62692]: raise self.value [ 741.779381] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 741.779381] env[62692]: updated_port = self._update_port( [ 741.779381] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 741.779381] env[62692]: _ensure_no_port_binding_failure(port) [ 741.779381] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 741.779381] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 741.783022] env[62692]: nova.exception.PortBindingFailed: Binding failed for port 42704cff-466f-442d-8911-a6365882ee2d, please check neutron logs for more information. [ 741.783022] env[62692]: Removing descriptor: 19 [ 741.784756] env[62692]: DEBUG nova.virt.hardware [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 741.785028] env[62692]: DEBUG nova.virt.hardware [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 741.785253] env[62692]: DEBUG nova.virt.hardware [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 741.785381] env[62692]: DEBUG nova.virt.hardware [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 741.785530] env[62692]: DEBUG nova.virt.hardware [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 741.786100] env[62692]: DEBUG nova.virt.hardware [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 741.786100] env[62692]: DEBUG nova.virt.hardware [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 741.786100] env[62692]: DEBUG nova.virt.hardware [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 741.786295] env[62692]: DEBUG nova.virt.hardware [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 741.786407] env[62692]: DEBUG nova.virt.hardware [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 741.786565] env[62692]: DEBUG nova.virt.hardware [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 741.787698] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09147274-c4aa-416a-9466-0646df1b66e9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.797911] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-735557a4-29bf-4fa5-855d-54935c012800 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.814233] env[62692]: ERROR nova.compute.manager [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 42704cff-466f-442d-8911-a6365882ee2d, please check neutron logs for more information. [ 741.814233] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] Traceback (most recent call last): [ 741.814233] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 741.814233] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] yield resources [ 741.814233] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 741.814233] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] self.driver.spawn(context, instance, image_meta, [ 741.814233] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 741.814233] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 741.814233] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 741.814233] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] vm_ref = self.build_virtual_machine(instance, [ 741.814233] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 741.814749] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] vif_infos = vmwarevif.get_vif_info(self._session, [ 741.814749] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 741.814749] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] for vif in network_info: [ 741.814749] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 741.814749] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] return self._sync_wrapper(fn, *args, **kwargs) [ 741.814749] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 741.814749] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] self.wait() [ 741.814749] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 741.814749] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] self[:] = self._gt.wait() [ 741.814749] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 741.814749] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] return self._exit_event.wait() [ 741.814749] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 741.814749] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] current.throw(*self._exc) [ 741.815299] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 741.815299] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] result = function(*args, **kwargs) [ 741.815299] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 741.815299] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] return func(*args, **kwargs) [ 741.815299] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 741.815299] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] raise e [ 741.815299] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 741.815299] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] nwinfo = self.network_api.allocate_for_instance( [ 741.815299] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 741.815299] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] created_port_ids = self._update_ports_for_instance( [ 741.815299] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 741.815299] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] with excutils.save_and_reraise_exception(): [ 741.815299] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 741.815913] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] self.force_reraise() [ 741.815913] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 741.815913] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] raise self.value [ 741.815913] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 741.815913] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] updated_port = self._update_port( [ 741.815913] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 741.815913] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] _ensure_no_port_binding_failure(port) [ 741.815913] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 741.815913] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] raise exception.PortBindingFailed(port_id=port['id']) [ 741.815913] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] nova.exception.PortBindingFailed: Binding failed for port 42704cff-466f-442d-8911-a6365882ee2d, please check neutron logs for more information. [ 741.815913] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] [ 741.815913] env[62692]: INFO nova.compute.manager [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Terminating instance [ 741.818207] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Acquiring lock "refresh_cache-700d3723-f787-4770-a409-642cd8085e9c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.986302] env[62692]: DEBUG oslo_concurrency.lockutils [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.079133] env[62692]: DEBUG nova.network.neutron [req-d4d945b7-ee43-4f3c-94ee-4221bcf2b0fa req-2fb3c713-2c17-48eb-9665-dbe2537242b8 service nova] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 742.143194] env[62692]: DEBUG nova.network.neutron [req-d4d945b7-ee43-4f3c-94ee-4221bcf2b0fa req-2fb3c713-2c17-48eb-9665-dbe2537242b8 service nova] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.290432] env[62692]: DEBUG nova.network.neutron [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 742.377503] env[62692]: DEBUG nova.network.neutron [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.634166] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bc31821-0288-45a7-ba7d-5319dec2eaaa {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.641265] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5762d844-55ea-4cda-998b-3e32b8cc1867 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.646461] env[62692]: DEBUG oslo_concurrency.lockutils [req-d4d945b7-ee43-4f3c-94ee-4221bcf2b0fa req-2fb3c713-2c17-48eb-9665-dbe2537242b8 service nova] Releasing lock "refresh_cache-700d3723-f787-4770-a409-642cd8085e9c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.647482] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Acquired lock "refresh_cache-700d3723-f787-4770-a409-642cd8085e9c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.648014] env[62692]: DEBUG nova.network.neutron [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 742.683209] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7b4fbb2-dea7-4c10-8fab-e35db600c08b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.688566] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquiring lock "2ccb76cf-f641-4306-a137-fb2417285df9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.688796] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Lock "2ccb76cf-f641-4306-a137-fb2417285df9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.694262] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6df15a6-713b-4df5-b8a0-609859fdefde {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.708718] env[62692]: DEBUG nova.compute.provider_tree [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 742.711122] env[62692]: DEBUG nova.network.neutron [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 742.769593] env[62692]: DEBUG nova.network.neutron [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.880545] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Releasing lock "refresh_cache-dbd51115-6b41-4019-955d-16a52b534e84" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.880860] env[62692]: DEBUG nova.compute.manager [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 742.883027] env[62692]: DEBUG nova.compute.manager [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 742.883027] env[62692]: DEBUG nova.network.neutron [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 742.904710] env[62692]: DEBUG nova.network.neutron [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 743.214234] env[62692]: DEBUG nova.scheduler.client.report [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 743.272053] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Releasing lock "refresh_cache-700d3723-f787-4770-a409-642cd8085e9c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.272533] env[62692]: DEBUG nova.compute.manager [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 743.272730] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 743.273015] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-abab531e-f461-4d9e-bb80-5ba7fd6aee17 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.282104] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b12356d0-cdfe-48fa-a027-871d6c97508a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.303856] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 700d3723-f787-4770-a409-642cd8085e9c could not be found. [ 743.304096] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 743.304283] env[62692]: INFO nova.compute.manager [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Took 0.03 seconds to destroy the instance on the hypervisor. [ 743.304655] env[62692]: DEBUG oslo.service.loopingcall [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 743.304707] env[62692]: DEBUG nova.compute.manager [-] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 743.304804] env[62692]: DEBUG nova.network.neutron [-] [instance: 700d3723-f787-4770-a409-642cd8085e9c] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 743.319142] env[62692]: DEBUG nova.network.neutron [-] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 743.406666] env[62692]: DEBUG nova.network.neutron [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.590649] env[62692]: DEBUG nova.compute.manager [req-561c01cb-abaf-4cda-98e1-ba2af35621b0 req-dae59d01-a006-4696-8e92-b33d1da81360 service nova] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Received event network-vif-deleted-42704cff-466f-442d-8911-a6365882ee2d {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 743.724228] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.962s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.724228] env[62692]: ERROR nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d9b44847-736e-4744-9caa-0c5572ecf152, please check neutron logs for more information. [ 743.724228] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Traceback (most recent call last): [ 743.724228] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 743.724228] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] self.driver.spawn(context, instance, image_meta, [ 743.724228] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 743.724228] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 743.724228] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 743.724228] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] vm_ref = self.build_virtual_machine(instance, [ 743.724490] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 743.724490] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] vif_infos = vmwarevif.get_vif_info(self._session, [ 743.724490] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 743.724490] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] for vif in network_info: [ 743.724490] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 743.724490] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] return self._sync_wrapper(fn, *args, **kwargs) [ 743.724490] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 743.724490] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] self.wait() [ 743.724490] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 743.724490] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] self[:] = self._gt.wait() [ 743.724490] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 743.724490] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] return self._exit_event.wait() [ 743.724490] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 743.724795] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] current.throw(*self._exc) [ 743.724795] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 743.724795] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] result = function(*args, **kwargs) [ 743.724795] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 743.724795] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] return func(*args, **kwargs) [ 743.724795] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 743.724795] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] raise e [ 743.724795] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.724795] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] nwinfo = self.network_api.allocate_for_instance( [ 743.724795] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 743.724795] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] created_port_ids = self._update_ports_for_instance( [ 743.724795] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 743.724795] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] with excutils.save_and_reraise_exception(): [ 743.725133] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.725133] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] self.force_reraise() [ 743.725133] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.725133] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] raise self.value [ 743.725133] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 743.725133] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] updated_port = self._update_port( [ 743.725133] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.725133] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] _ensure_no_port_binding_failure(port) [ 743.725133] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.725133] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] raise exception.PortBindingFailed(port_id=port['id']) [ 743.725133] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] nova.exception.PortBindingFailed: Binding failed for port d9b44847-736e-4744-9caa-0c5572ecf152, please check neutron logs for more information. [ 743.725133] env[62692]: ERROR nova.compute.manager [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] [ 743.725417] env[62692]: DEBUG nova.compute.utils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Binding failed for port d9b44847-736e-4744-9caa-0c5572ecf152, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 743.725417] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.584s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.726735] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Build of instance 49d69437-7046-4cb2-914c-8544dca70d3f was re-scheduled: Binding failed for port d9b44847-736e-4744-9caa-0c5572ecf152, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 743.727171] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 743.727407] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Acquiring lock "refresh_cache-49d69437-7046-4cb2-914c-8544dca70d3f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.727552] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Acquired lock "refresh_cache-49d69437-7046-4cb2-914c-8544dca70d3f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.727711] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 743.821838] env[62692]: DEBUG nova.network.neutron [-] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.911693] env[62692]: INFO nova.compute.manager [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: dbd51115-6b41-4019-955d-16a52b534e84] Took 1.03 seconds to deallocate network for instance. [ 744.255132] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 744.330808] env[62692]: INFO nova.compute.manager [-] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Took 1.03 seconds to deallocate network for instance. [ 744.334033] env[62692]: DEBUG nova.compute.claims [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 744.334033] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.379938] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.674416] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30daefb5-8682-44c4-930a-2c3137d13d07 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.683010] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fec121f8-f0ef-41c7-b102-e1456553fc97 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.713241] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccfbf640-30b8-4e1b-bd6e-5c2bba0f66e7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.721763] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb9037cc-7f2f-4666-ad05-311553473753 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.736988] env[62692]: DEBUG nova.compute.provider_tree [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 744.882961] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Releasing lock "refresh_cache-49d69437-7046-4cb2-914c-8544dca70d3f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.883210] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 744.883343] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 744.883516] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 744.899346] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 744.937290] env[62692]: INFO nova.scheduler.client.report [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Deleted allocations for instance dbd51115-6b41-4019-955d-16a52b534e84 [ 745.233363] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Acquiring lock "c6a965d3-1c52-4f95-a226-9d15b7197ce7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.233481] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Lock "c6a965d3-1c52-4f95-a226-9d15b7197ce7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.240780] env[62692]: DEBUG nova.scheduler.client.report [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 745.402258] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.451023] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7339506a-0ca1-4864-ac36-f2ec06c35e90 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "dbd51115-6b41-4019-955d-16a52b534e84" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.415s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.747584] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.022s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.747584] env[62692]: ERROR nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3259562d-0513-42ef-8c6a-6e9c0e41f58c, please check neutron logs for more information. [ 745.747584] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Traceback (most recent call last): [ 745.747584] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 745.747584] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] self.driver.spawn(context, instance, image_meta, [ 745.747584] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 745.747584] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 745.747584] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 745.747584] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] vm_ref = self.build_virtual_machine(instance, [ 745.747847] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 745.747847] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] vif_infos = vmwarevif.get_vif_info(self._session, [ 745.747847] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 745.747847] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] for vif in network_info: [ 745.747847] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 745.747847] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] return self._sync_wrapper(fn, *args, **kwargs) [ 745.747847] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 745.747847] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] self.wait() [ 745.747847] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 745.747847] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] self[:] = self._gt.wait() [ 745.747847] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 745.747847] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] return self._exit_event.wait() [ 745.747847] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 745.748208] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] current.throw(*self._exc) [ 745.748208] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 745.748208] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] result = function(*args, **kwargs) [ 745.748208] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 745.748208] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] return func(*args, **kwargs) [ 745.748208] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 745.748208] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] raise e [ 745.748208] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 745.748208] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] nwinfo = self.network_api.allocate_for_instance( [ 745.748208] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 745.748208] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] created_port_ids = self._update_ports_for_instance( [ 745.748208] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 745.748208] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] with excutils.save_and_reraise_exception(): [ 745.748710] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 745.748710] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] self.force_reraise() [ 745.748710] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 745.748710] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] raise self.value [ 745.748710] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 745.748710] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] updated_port = self._update_port( [ 745.748710] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 745.748710] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] _ensure_no_port_binding_failure(port) [ 745.748710] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 745.748710] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] raise exception.PortBindingFailed(port_id=port['id']) [ 745.748710] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] nova.exception.PortBindingFailed: Binding failed for port 3259562d-0513-42ef-8c6a-6e9c0e41f58c, please check neutron logs for more information. [ 745.748710] env[62692]: ERROR nova.compute.manager [instance: db727822-af78-49bc-98d9-9caabc35b3b6] [ 745.749158] env[62692]: DEBUG nova.compute.utils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Binding failed for port 3259562d-0513-42ef-8c6a-6e9c0e41f58c, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 745.750058] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.149s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.752551] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Build of instance db727822-af78-49bc-98d9-9caabc35b3b6 was re-scheduled: Binding failed for port 3259562d-0513-42ef-8c6a-6e9c0e41f58c, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 745.753174] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 745.753465] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Acquiring lock "refresh_cache-db727822-af78-49bc-98d9-9caabc35b3b6" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.753668] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Acquired lock "refresh_cache-db727822-af78-49bc-98d9-9caabc35b3b6" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.753895] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 745.905041] env[62692]: INFO nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 49d69437-7046-4cb2-914c-8544dca70d3f] Took 1.02 seconds to deallocate network for instance. [ 745.952385] env[62692]: DEBUG nova.compute.manager [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 746.286094] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 746.387797] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.471700] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.629590] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c9d3cb1-9762-4038-871c-f21ad5077815 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.637992] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7f78518-fc2b-4ed3-8f4c-8e8ed0364517 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.670783] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3979c639-ddb6-461f-9806-2fe0f9949379 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.678418] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bce7e86-b9aa-45be-ace2-748d29790cac {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.691918] env[62692]: DEBUG nova.compute.provider_tree [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 746.890873] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Releasing lock "refresh_cache-db727822-af78-49bc-98d9-9caabc35b3b6" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.891138] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 746.891632] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 746.891863] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 746.907477] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 746.937956] env[62692]: INFO nova.scheduler.client.report [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Deleted allocations for instance 49d69437-7046-4cb2-914c-8544dca70d3f [ 747.195056] env[62692]: DEBUG nova.scheduler.client.report [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 747.287773] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "64a8329d-e0e9-4c2d-bd1f-844ee40eb980" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.288101] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "64a8329d-e0e9-4c2d-bd1f-844ee40eb980" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.410619] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.447519] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Lock "49d69437-7046-4cb2-914c-8544dca70d3f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.766s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.700155] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.951s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.700866] env[62692]: ERROR nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8fcbd7c0-2426-4395-96f4-617b331079ab, please check neutron logs for more information. [ 747.700866] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Traceback (most recent call last): [ 747.700866] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 747.700866] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] self.driver.spawn(context, instance, image_meta, [ 747.700866] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 747.700866] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] self._vmops.spawn(context, instance, image_meta, injected_files, [ 747.700866] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 747.700866] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] vm_ref = self.build_virtual_machine(instance, [ 747.700866] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 747.700866] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] vif_infos = vmwarevif.get_vif_info(self._session, [ 747.700866] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 747.701206] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] for vif in network_info: [ 747.701206] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 747.701206] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] return self._sync_wrapper(fn, *args, **kwargs) [ 747.701206] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 747.701206] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] self.wait() [ 747.701206] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 747.701206] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] self[:] = self._gt.wait() [ 747.701206] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 747.701206] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] return self._exit_event.wait() [ 747.701206] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 747.701206] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] current.throw(*self._exc) [ 747.701206] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 747.701206] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] result = function(*args, **kwargs) [ 747.701517] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 747.701517] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] return func(*args, **kwargs) [ 747.701517] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 747.701517] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] raise e [ 747.701517] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 747.701517] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] nwinfo = self.network_api.allocate_for_instance( [ 747.701517] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 747.701517] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] created_port_ids = self._update_ports_for_instance( [ 747.701517] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 747.701517] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] with excutils.save_and_reraise_exception(): [ 747.701517] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.701517] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] self.force_reraise() [ 747.701517] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.701829] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] raise self.value [ 747.701829] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 747.701829] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] updated_port = self._update_port( [ 747.701829] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.701829] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] _ensure_no_port_binding_failure(port) [ 747.701829] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.701829] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] raise exception.PortBindingFailed(port_id=port['id']) [ 747.701829] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] nova.exception.PortBindingFailed: Binding failed for port 8fcbd7c0-2426-4395-96f4-617b331079ab, please check neutron logs for more information. [ 747.701829] env[62692]: ERROR nova.compute.manager [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] [ 747.701829] env[62692]: DEBUG nova.compute.utils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Binding failed for port 8fcbd7c0-2426-4395-96f4-617b331079ab, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 747.702793] env[62692]: DEBUG oslo_concurrency.lockutils [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.639s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.704299] env[62692]: INFO nova.compute.claims [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 747.707027] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Build of instance 0266b67a-7c53-471d-a167-ac7362c4b9be was re-scheduled: Binding failed for port 8fcbd7c0-2426-4395-96f4-617b331079ab, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 747.707446] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 747.707666] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Acquiring lock "refresh_cache-0266b67a-7c53-471d-a167-ac7362c4b9be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.707811] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Acquired lock "refresh_cache-0266b67a-7c53-471d-a167-ac7362c4b9be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.707968] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 747.914046] env[62692]: INFO nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: db727822-af78-49bc-98d9-9caabc35b3b6] Took 1.02 seconds to deallocate network for instance. [ 747.949898] env[62692]: DEBUG nova.compute.manager [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 748.231156] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 748.312626] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.470037] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.817643] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Releasing lock "refresh_cache-0266b67a-7c53-471d-a167-ac7362c4b9be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.817643] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 748.817643] env[62692]: DEBUG nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 748.817643] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 748.837016] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 748.937313] env[62692]: INFO nova.scheduler.client.report [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Deleted allocations for instance db727822-af78-49bc-98d9-9caabc35b3b6 [ 749.062084] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26bb49ab-b01a-4e46-b4ff-c3be646e6205 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.070015] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e0b877e-489c-4869-bd6d-3d765a2e9674 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.099970] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fe35773-6bd2-4b0b-ae73-895dff748d05 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.107436] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dd92d3b-2812-4757-afec-3e9cb0f61a2d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.122181] env[62692]: DEBUG nova.compute.provider_tree [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 749.340514] env[62692]: DEBUG nova.network.neutron [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.448659] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Lock "db727822-af78-49bc-98d9-9caabc35b3b6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.730s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.625595] env[62692]: DEBUG nova.scheduler.client.report [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 749.843048] env[62692]: INFO nova.compute.manager [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] [instance: 0266b67a-7c53-471d-a167-ac7362c4b9be] Took 1.03 seconds to deallocate network for instance. [ 749.950992] env[62692]: DEBUG nova.compute.manager [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 750.130429] env[62692]: DEBUG oslo_concurrency.lockutils [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.427s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.131088] env[62692]: DEBUG nova.compute.manager [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 750.133987] env[62692]: DEBUG oslo_concurrency.lockutils [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.139s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.470525] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.639366] env[62692]: DEBUG nova.compute.utils [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 750.644042] env[62692]: DEBUG nova.compute.manager [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 750.644390] env[62692]: DEBUG nova.network.neutron [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 750.687644] env[62692]: DEBUG nova.policy [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b053dac28bfb4e13949f57d319e6983a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c1f6f76c8c064933a0267c737542397b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 750.873974] env[62692]: INFO nova.scheduler.client.report [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Deleted allocations for instance 0266b67a-7c53-471d-a167-ac7362c4b9be [ 750.989704] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f463ba-40f6-4b68-a1a1-b9507b14a2b5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.997231] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-324a1a56-a0fd-4803-9b15-333295111b09 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.027905] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e1817ef-851f-421f-a455-d84b2518aa04 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.034757] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-821b771f-5ea5-49cb-b445-c34c3a0f2202 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.047499] env[62692]: DEBUG nova.compute.provider_tree [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 751.144947] env[62692]: DEBUG nova.compute.manager [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 751.209963] env[62692]: DEBUG nova.network.neutron [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Successfully created port: 8f492038-90fc-40ae-9f44-214df7488a50 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 751.386216] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a2b6e5b0-3f11-40a2-ab8e-00925b480cea tempest-ListServersNegativeTestJSON-1780715791 tempest-ListServersNegativeTestJSON-1780715791-project-member] Lock "0266b67a-7c53-471d-a167-ac7362c4b9be" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.639s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.550435] env[62692]: DEBUG nova.scheduler.client.report [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 751.888806] env[62692]: DEBUG nova.compute.manager [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 752.058414] env[62692]: DEBUG oslo_concurrency.lockutils [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.924s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.059067] env[62692]: ERROR nova.compute.manager [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ec8028c0-e992-442a-a555-ea6f3e790f35, please check neutron logs for more information. [ 752.059067] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Traceback (most recent call last): [ 752.059067] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 752.059067] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] self.driver.spawn(context, instance, image_meta, [ 752.059067] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 752.059067] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] self._vmops.spawn(context, instance, image_meta, injected_files, [ 752.059067] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 752.059067] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] vm_ref = self.build_virtual_machine(instance, [ 752.059067] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 752.059067] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] vif_infos = vmwarevif.get_vif_info(self._session, [ 752.059067] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 752.059388] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] for vif in network_info: [ 752.059388] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 752.059388] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] return self._sync_wrapper(fn, *args, **kwargs) [ 752.059388] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 752.059388] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] self.wait() [ 752.059388] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 752.059388] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] self[:] = self._gt.wait() [ 752.059388] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 752.059388] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] return self._exit_event.wait() [ 752.059388] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 752.059388] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] current.throw(*self._exc) [ 752.059388] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 752.059388] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] result = function(*args, **kwargs) [ 752.059695] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 752.059695] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] return func(*args, **kwargs) [ 752.059695] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 752.059695] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] raise e [ 752.059695] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 752.059695] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] nwinfo = self.network_api.allocate_for_instance( [ 752.059695] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 752.059695] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] created_port_ids = self._update_ports_for_instance( [ 752.059695] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 752.059695] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] with excutils.save_and_reraise_exception(): [ 752.059695] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 752.059695] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] self.force_reraise() [ 752.059695] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 752.060015] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] raise self.value [ 752.060015] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 752.060015] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] updated_port = self._update_port( [ 752.060015] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 752.060015] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] _ensure_no_port_binding_failure(port) [ 752.060015] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 752.060015] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] raise exception.PortBindingFailed(port_id=port['id']) [ 752.060015] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] nova.exception.PortBindingFailed: Binding failed for port ec8028c0-e992-442a-a555-ea6f3e790f35, please check neutron logs for more information. [ 752.060015] env[62692]: ERROR nova.compute.manager [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] [ 752.060015] env[62692]: DEBUG nova.compute.utils [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Binding failed for port ec8028c0-e992-442a-a555-ea6f3e790f35, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 752.060979] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 14.688s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.062968] env[62692]: DEBUG nova.compute.manager [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Build of instance a5fa20ab-66b7-4632-8a80-ac031263c042 was re-scheduled: Binding failed for port ec8028c0-e992-442a-a555-ea6f3e790f35, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 752.063499] env[62692]: DEBUG nova.compute.manager [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 752.063845] env[62692]: DEBUG oslo_concurrency.lockutils [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Acquiring lock "refresh_cache-a5fa20ab-66b7-4632-8a80-ac031263c042" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.064077] env[62692]: DEBUG oslo_concurrency.lockutils [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Acquired lock "refresh_cache-a5fa20ab-66b7-4632-8a80-ac031263c042" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.064300] env[62692]: DEBUG nova.network.neutron [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 752.117784] env[62692]: DEBUG nova.compute.manager [req-a37e5f66-6a18-4204-8b48-d0547e88fc1b req-13a77c76-6565-4089-9667-415b2bb44d6d service nova] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Received event network-changed-8f492038-90fc-40ae-9f44-214df7488a50 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 752.118011] env[62692]: DEBUG nova.compute.manager [req-a37e5f66-6a18-4204-8b48-d0547e88fc1b req-13a77c76-6565-4089-9667-415b2bb44d6d service nova] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Refreshing instance network info cache due to event network-changed-8f492038-90fc-40ae-9f44-214df7488a50. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 752.118252] env[62692]: DEBUG oslo_concurrency.lockutils [req-a37e5f66-6a18-4204-8b48-d0547e88fc1b req-13a77c76-6565-4089-9667-415b2bb44d6d service nova] Acquiring lock "refresh_cache-49ee893a-313e-42ed-b932-352d450e8645" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.118405] env[62692]: DEBUG oslo_concurrency.lockutils [req-a37e5f66-6a18-4204-8b48-d0547e88fc1b req-13a77c76-6565-4089-9667-415b2bb44d6d service nova] Acquired lock "refresh_cache-49ee893a-313e-42ed-b932-352d450e8645" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.118565] env[62692]: DEBUG nova.network.neutron [req-a37e5f66-6a18-4204-8b48-d0547e88fc1b req-13a77c76-6565-4089-9667-415b2bb44d6d service nova] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Refreshing network info cache for port 8f492038-90fc-40ae-9f44-214df7488a50 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 752.155235] env[62692]: DEBUG nova.compute.manager [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 752.179705] env[62692]: DEBUG nova.virt.hardware [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 752.179705] env[62692]: DEBUG nova.virt.hardware [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 752.179899] env[62692]: DEBUG nova.virt.hardware [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 752.179988] env[62692]: DEBUG nova.virt.hardware [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 752.180154] env[62692]: DEBUG nova.virt.hardware [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 752.180296] env[62692]: DEBUG nova.virt.hardware [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 752.180504] env[62692]: DEBUG nova.virt.hardware [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 752.180749] env[62692]: DEBUG nova.virt.hardware [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 752.180842] env[62692]: DEBUG nova.virt.hardware [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 752.181264] env[62692]: DEBUG nova.virt.hardware [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 752.181264] env[62692]: DEBUG nova.virt.hardware [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 752.181996] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1092ad2d-3a29-44ee-9ea4-a9fcdef42c02 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.192921] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e97244a-76bc-4a82-a4fa-ce3d60d65132 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.230341] env[62692]: ERROR nova.compute.manager [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8f492038-90fc-40ae-9f44-214df7488a50, please check neutron logs for more information. [ 752.230341] env[62692]: ERROR nova.compute.manager Traceback (most recent call last): [ 752.230341] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 752.230341] env[62692]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 752.230341] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 752.230341] env[62692]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 752.230341] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 752.230341] env[62692]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 752.230341] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 752.230341] env[62692]: ERROR nova.compute.manager self.force_reraise() [ 752.230341] env[62692]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 752.230341] env[62692]: ERROR nova.compute.manager raise self.value [ 752.230341] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 752.230341] env[62692]: ERROR nova.compute.manager updated_port = self._update_port( [ 752.230341] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 752.230341] env[62692]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 752.230906] env[62692]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 752.230906] env[62692]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 752.230906] env[62692]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8f492038-90fc-40ae-9f44-214df7488a50, please check neutron logs for more information. [ 752.230906] env[62692]: ERROR nova.compute.manager [ 752.230906] env[62692]: Traceback (most recent call last): [ 752.230906] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 752.230906] env[62692]: listener.cb(fileno) [ 752.230906] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 752.230906] env[62692]: result = function(*args, **kwargs) [ 752.230906] env[62692]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 752.230906] env[62692]: return func(*args, **kwargs) [ 752.230906] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 752.230906] env[62692]: raise e [ 752.230906] env[62692]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 752.230906] env[62692]: nwinfo = self.network_api.allocate_for_instance( [ 752.230906] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 752.230906] env[62692]: created_port_ids = self._update_ports_for_instance( [ 752.230906] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 752.230906] env[62692]: with excutils.save_and_reraise_exception(): [ 752.230906] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 752.230906] env[62692]: self.force_reraise() [ 752.230906] env[62692]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 752.230906] env[62692]: raise self.value [ 752.230906] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 752.230906] env[62692]: updated_port = self._update_port( [ 752.230906] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 752.230906] env[62692]: _ensure_no_port_binding_failure(port) [ 752.230906] env[62692]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 752.230906] env[62692]: raise exception.PortBindingFailed(port_id=port['id']) [ 752.231733] env[62692]: nova.exception.PortBindingFailed: Binding failed for port 8f492038-90fc-40ae-9f44-214df7488a50, please check neutron logs for more information. [ 752.231733] env[62692]: Removing descriptor: 19 [ 752.231733] env[62692]: ERROR nova.compute.manager [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8f492038-90fc-40ae-9f44-214df7488a50, please check neutron logs for more information. [ 752.231733] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] Traceback (most recent call last): [ 752.231733] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 752.231733] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] yield resources [ 752.231733] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 752.231733] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] self.driver.spawn(context, instance, image_meta, [ 752.231733] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 752.231733] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] self._vmops.spawn(context, instance, image_meta, injected_files, [ 752.231733] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 752.231733] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] vm_ref = self.build_virtual_machine(instance, [ 752.232052] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 752.232052] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] vif_infos = vmwarevif.get_vif_info(self._session, [ 752.232052] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 752.232052] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] for vif in network_info: [ 752.232052] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 752.232052] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] return self._sync_wrapper(fn, *args, **kwargs) [ 752.232052] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 752.232052] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] self.wait() [ 752.232052] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 752.232052] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] self[:] = self._gt.wait() [ 752.232052] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 752.232052] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] return self._exit_event.wait() [ 752.232052] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 752.232399] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] result = hub.switch() [ 752.232399] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 752.232399] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] return self.greenlet.switch() [ 752.232399] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 752.232399] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] result = function(*args, **kwargs) [ 752.232399] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 752.232399] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] return func(*args, **kwargs) [ 752.232399] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 752.232399] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] raise e [ 752.232399] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 752.232399] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] nwinfo = self.network_api.allocate_for_instance( [ 752.232399] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 752.232399] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] created_port_ids = self._update_ports_for_instance( [ 752.232749] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 752.232749] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] with excutils.save_and_reraise_exception(): [ 752.232749] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 752.232749] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] self.force_reraise() [ 752.232749] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 752.232749] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] raise self.value [ 752.232749] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 752.232749] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] updated_port = self._update_port( [ 752.232749] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 752.232749] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] _ensure_no_port_binding_failure(port) [ 752.232749] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 752.232749] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] raise exception.PortBindingFailed(port_id=port['id']) [ 752.233071] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] nova.exception.PortBindingFailed: Binding failed for port 8f492038-90fc-40ae-9f44-214df7488a50, please check neutron logs for more information. [ 752.233071] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] [ 752.233071] env[62692]: INFO nova.compute.manager [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Terminating instance [ 752.233722] env[62692]: DEBUG oslo_concurrency.lockutils [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Acquiring lock "refresh_cache-49ee893a-313e-42ed-b932-352d450e8645" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.414071] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.586073] env[62692]: DEBUG nova.network.neutron [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 752.642875] env[62692]: DEBUG nova.network.neutron [req-a37e5f66-6a18-4204-8b48-d0547e88fc1b req-13a77c76-6565-4089-9667-415b2bb44d6d service nova] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 752.810291] env[62692]: DEBUG nova.network.neutron [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.840567] env[62692]: DEBUG nova.network.neutron [req-a37e5f66-6a18-4204-8b48-d0547e88fc1b req-13a77c76-6565-4089-9667-415b2bb44d6d service nova] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.313855] env[62692]: DEBUG oslo_concurrency.lockutils [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Releasing lock "refresh_cache-a5fa20ab-66b7-4632-8a80-ac031263c042" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.314179] env[62692]: DEBUG nova.compute.manager [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 753.314303] env[62692]: DEBUG nova.compute.manager [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 753.314475] env[62692]: DEBUG nova.network.neutron [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 753.337388] env[62692]: DEBUG nova.network.neutron [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 753.343669] env[62692]: DEBUG oslo_concurrency.lockutils [req-a37e5f66-6a18-4204-8b48-d0547e88fc1b req-13a77c76-6565-4089-9667-415b2bb44d6d service nova] Releasing lock "refresh_cache-49ee893a-313e-42ed-b932-352d450e8645" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.344056] env[62692]: DEBUG oslo_concurrency.lockutils [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Acquired lock "refresh_cache-49ee893a-313e-42ed-b932-352d450e8645" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.344247] env[62692]: DEBUG nova.network.neutron [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 753.601945] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance a5fa20ab-66b7-4632-8a80-ac031263c042 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 753.601945] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance ed43eb23-0f75-4a8b-96a1-142e1abe400f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 753.601945] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 700d3723-f787-4770-a409-642cd8085e9c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 753.602232] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 49ee893a-313e-42ed-b932-352d450e8645 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 753.840240] env[62692]: DEBUG nova.network.neutron [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.870695] env[62692]: DEBUG nova.network.neutron [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 753.977944] env[62692]: DEBUG nova.network.neutron [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.104714] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance d99737d8-2eb0-40ee-b61e-6c736c84ea59 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 754.148329] env[62692]: DEBUG nova.compute.manager [req-82249dfa-6b23-4c4c-9f5c-869bbec8e73a req-829e913b-7b18-4fd1-bb7d-c2e259b5d577 service nova] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Received event network-vif-deleted-8f492038-90fc-40ae-9f44-214df7488a50 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 754.343624] env[62692]: INFO nova.compute.manager [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] [instance: a5fa20ab-66b7-4632-8a80-ac031263c042] Took 1.03 seconds to deallocate network for instance. [ 754.480476] env[62692]: DEBUG oslo_concurrency.lockutils [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Releasing lock "refresh_cache-49ee893a-313e-42ed-b932-352d450e8645" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.481157] env[62692]: DEBUG nova.compute.manager [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 754.481946] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 754.482288] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2d3d7c9a-a786-4228-a43c-a4f59965fdfb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.492510] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21a1af80-d55c-40ca-afd3-9a142183f352 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.520977] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 49ee893a-313e-42ed-b932-352d450e8645 could not be found. [ 754.520977] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 754.520977] env[62692]: INFO nova.compute.manager [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Took 0.04 seconds to destroy the instance on the hypervisor. [ 754.520977] env[62692]: DEBUG oslo.service.loopingcall [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 754.520977] env[62692]: DEBUG nova.compute.manager [-] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 754.521244] env[62692]: DEBUG nova.network.neutron [-] [instance: 49ee893a-313e-42ed-b932-352d450e8645] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 754.545799] env[62692]: DEBUG nova.network.neutron [-] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 754.607180] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 628cc505-3edf-4066-91be-da009ebcf219 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 755.052057] env[62692]: DEBUG nova.network.neutron [-] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.109877] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance abee88ff-5cf5-4bf1-91e0-93b19cf30046 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 755.376826] env[62692]: INFO nova.scheduler.client.report [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Deleted allocations for instance a5fa20ab-66b7-4632-8a80-ac031263c042 [ 755.555259] env[62692]: INFO nova.compute.manager [-] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Took 1.03 seconds to deallocate network for instance. [ 755.561102] env[62692]: DEBUG nova.compute.claims [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Aborting claim: {{(pid=62692) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 755.561420] env[62692]: DEBUG oslo_concurrency.lockutils [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.614912] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance fa7bb883-0065-4ecc-9eb5-8c5443344a2a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 755.886833] env[62692]: DEBUG oslo_concurrency.lockutils [None req-92e1c3cc-cce2-4da5-9001-37ff609a4594 tempest-ImagesOneServerTestJSON-1781947144 tempest-ImagesOneServerTestJSON-1781947144-project-member] Lock "a5fa20ab-66b7-4632-8a80-ac031263c042" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.824s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.118764] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance c1818ed2-9e27-4738-bc17-98832318cf61 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 756.393611] env[62692]: DEBUG nova.compute.manager [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 756.621480] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 5a08fa8a-f9fe-4879-bb7b-baa04097df6c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 756.925887] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.124267] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 9b8e830c-61b7-4dd2-8324-d3a96eec1465 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 757.630555] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 105e6fd5-4eff-4a0c-9a4b-e6deade781d5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 758.135235] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance bb547773-d176-4c8e-a0fa-a374d5050b1b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 758.639530] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 759.144596] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 6097d6b2-52d5-4765-94c0-d5f3609464d8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 759.647157] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 27583cb4-0478-44d8-8970-478682271513 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 760.153312] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance aa0666ce-4d4f-4333-8acb-ea0cfb438cd9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 760.657612] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 1848c251-0a0c-4c36-8bd0-ff37befde2a0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 761.042354] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Acquiring lock "18eb004f-a16b-477a-8d05-ed23a418b069" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.042757] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Lock "18eb004f-a16b-477a-8d05-ed23a418b069" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.164249] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 81ce340c-fbef-4932-983c-595843530dbc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 761.668386] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 762.174796] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 7a3f9c7d-03cf-4177-88db-acf33cc03feb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 762.677929] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance b1d6409a-2733-470f-a929-672fe1631b1b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 763.172125] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "e2383a6a-3581-40fc-a0eb-6981acdbf54a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.172365] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "e2383a6a-3581-40fc-a0eb-6981acdbf54a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.180827] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 763.683599] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 8b75e781-930e-4885-81d2-8a8929d6c39b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 764.185812] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 2ccb76cf-f641-4306-a137-fb2417285df9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 764.691029] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance c6a965d3-1c52-4f95-a226-9d15b7197ce7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 765.193026] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 64a8329d-e0e9-4c2d-bd1f-844ee40eb980 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 765.193273] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62692) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 765.193428] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62692) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 765.507351] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58285db1-fd6c-42ac-b26d-8173a0de7af0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.515193] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ee764b3-0977-448a-832c-6c9b66773a27 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.545029] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7c3c2b7-ece9-4316-8a00-7fb75a7f36b2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.552819] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-387c1bcd-8b81-4988-9180-a678045ee670 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.566263] env[62692]: DEBUG nova.compute.provider_tree [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 766.069910] env[62692]: DEBUG nova.scheduler.client.report [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 766.574468] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62692) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 766.574809] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 14.514s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.575132] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.952s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.433114] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d0ca782-728f-4375-82bb-d76c48300eef {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.440559] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59c58c82-3956-4730-825b-b72bf647cb1e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.470066] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cf28a0a-20b1-48e3-93e2-7e6a7f3e9e45 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.477228] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f04081b5-82db-46e3-9d20-0e8dd01aa417 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.489850] env[62692]: DEBUG nova.compute.provider_tree [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.995298] env[62692]: DEBUG nova.scheduler.client.report [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 768.500309] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.925s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.501089] env[62692]: ERROR nova.compute.manager [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3736cd8e-be04-4c79-985a-54953ea3d16c, please check neutron logs for more information. [ 768.501089] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Traceback (most recent call last): [ 768.501089] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 768.501089] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] self.driver.spawn(context, instance, image_meta, [ 768.501089] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 768.501089] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 768.501089] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 768.501089] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] vm_ref = self.build_virtual_machine(instance, [ 768.501089] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 768.501089] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] vif_infos = vmwarevif.get_vif_info(self._session, [ 768.501089] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 768.501532] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] for vif in network_info: [ 768.501532] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 768.501532] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] return self._sync_wrapper(fn, *args, **kwargs) [ 768.501532] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 768.501532] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] self.wait() [ 768.501532] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 768.501532] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] self[:] = self._gt.wait() [ 768.501532] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 768.501532] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] return self._exit_event.wait() [ 768.501532] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 768.501532] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] current.throw(*self._exc) [ 768.501532] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 768.501532] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] result = function(*args, **kwargs) [ 768.501915] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 768.501915] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] return func(*args, **kwargs) [ 768.501915] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 768.501915] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] raise e [ 768.501915] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 768.501915] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] nwinfo = self.network_api.allocate_for_instance( [ 768.501915] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 768.501915] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] created_port_ids = self._update_ports_for_instance( [ 768.501915] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 768.501915] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] with excutils.save_and_reraise_exception(): [ 768.501915] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 768.501915] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] self.force_reraise() [ 768.501915] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 768.502250] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] raise self.value [ 768.502250] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 768.502250] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] updated_port = self._update_port( [ 768.502250] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 768.502250] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] _ensure_no_port_binding_failure(port) [ 768.502250] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 768.502250] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] raise exception.PortBindingFailed(port_id=port['id']) [ 768.502250] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] nova.exception.PortBindingFailed: Binding failed for port 3736cd8e-be04-4c79-985a-54953ea3d16c, please check neutron logs for more information. [ 768.502250] env[62692]: ERROR nova.compute.manager [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] [ 768.502250] env[62692]: DEBUG nova.compute.utils [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Binding failed for port 3736cd8e-be04-4c79-985a-54953ea3d16c, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 768.503331] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.905s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.504749] env[62692]: INFO nova.compute.claims [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 768.507649] env[62692]: DEBUG nova.compute.manager [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Build of instance ed43eb23-0f75-4a8b-96a1-142e1abe400f was re-scheduled: Binding failed for port 3736cd8e-be04-4c79-985a-54953ea3d16c, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 768.508079] env[62692]: DEBUG nova.compute.manager [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 768.508394] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Acquiring lock "refresh_cache-ed43eb23-0f75-4a8b-96a1-142e1abe400f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.508744] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Acquired lock "refresh_cache-ed43eb23-0f75-4a8b-96a1-142e1abe400f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.509048] env[62692]: DEBUG nova.network.neutron [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 769.035318] env[62692]: DEBUG nova.network.neutron [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 769.124030] env[62692]: DEBUG nova.network.neutron [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.628379] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Releasing lock "refresh_cache-ed43eb23-0f75-4a8b-96a1-142e1abe400f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.628379] env[62692]: DEBUG nova.compute.manager [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 769.628379] env[62692]: DEBUG nova.compute.manager [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 769.628379] env[62692]: DEBUG nova.network.neutron [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 769.646326] env[62692]: DEBUG nova.network.neutron [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 769.842784] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-380d42d9-3233-4aef-8698-5b6718ff4521 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.849885] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6999fecc-e36c-4601-9041-7f924f5b5f79 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.880175] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f61ef4d2-afa8-4832-995c-803af855d259 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.887187] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65725744-95b4-4c0b-9a9f-be68eefce4a7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.900031] env[62692]: DEBUG nova.compute.provider_tree [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 770.149702] env[62692]: DEBUG nova.network.neutron [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.402978] env[62692]: DEBUG nova.scheduler.client.report [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 770.653047] env[62692]: INFO nova.compute.manager [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] [instance: ed43eb23-0f75-4a8b-96a1-142e1abe400f] Took 1.02 seconds to deallocate network for instance. [ 770.908220] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.405s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.908903] env[62692]: DEBUG nova.compute.manager [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 770.911655] env[62692]: DEBUG oslo_concurrency.lockutils [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.926s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.913145] env[62692]: INFO nova.compute.claims [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 771.417846] env[62692]: DEBUG nova.compute.utils [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 771.421136] env[62692]: DEBUG nova.compute.manager [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 771.421314] env[62692]: DEBUG nova.network.neutron [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 771.472752] env[62692]: DEBUG nova.policy [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6e9cd6d33e744da4a2f678226b684845', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1037e5d436bd429391159b87a712f6e1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 771.683160] env[62692]: INFO nova.scheduler.client.report [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Deleted allocations for instance ed43eb23-0f75-4a8b-96a1-142e1abe400f [ 771.814134] env[62692]: DEBUG nova.network.neutron [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Successfully created port: 7f2cabdf-cbda-4cf7-a518-13f619716d12 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 771.925179] env[62692]: DEBUG nova.compute.manager [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 772.191840] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cda2048c-4b55-4e1f-b807-0cfd42ac2cc9 tempest-ServersTestManualDisk-1096614721 tempest-ServersTestManualDisk-1096614721-project-member] Lock "ed43eb23-0f75-4a8b-96a1-142e1abe400f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.967s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.315329] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8c4041f-c4a9-4b0d-800c-dd7d9981921d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.327198] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7e75466-2fe8-425e-978f-c78e9d4c27c9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.353274] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6becb6a-f514-4c91-ac2e-cfeb9c910ccc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.360361] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13911ce4-737c-4ec0-aa75-2a9a826546f2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.374261] env[62692]: DEBUG nova.compute.provider_tree [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 772.695765] env[62692]: DEBUG nova.compute.manager [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 772.877298] env[62692]: DEBUG nova.scheduler.client.report [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 772.936246] env[62692]: DEBUG nova.compute.manager [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 772.963185] env[62692]: DEBUG nova.virt.hardware [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 772.963360] env[62692]: DEBUG nova.virt.hardware [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 772.963519] env[62692]: DEBUG nova.virt.hardware [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 772.963701] env[62692]: DEBUG nova.virt.hardware [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 772.963844] env[62692]: DEBUG nova.virt.hardware [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 772.963987] env[62692]: DEBUG nova.virt.hardware [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 772.964214] env[62692]: DEBUG nova.virt.hardware [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 772.964372] env[62692]: DEBUG nova.virt.hardware [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 772.964538] env[62692]: DEBUG nova.virt.hardware [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 772.964692] env[62692]: DEBUG nova.virt.hardware [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 772.964858] env[62692]: DEBUG nova.virt.hardware [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 772.965740] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5339766e-9de7-4b68-9f92-2f3097c2f961 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.974049] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e382c9d6-b58d-48e2-938d-c5e19b94bc5f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.223223] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.383386] env[62692]: DEBUG nova.compute.manager [req-1a419b02-d71b-4d49-a6cf-73374a62d44d req-ea0f237f-1888-4aea-804c-e21bb6a3b97f service nova] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Received event network-vif-plugged-7f2cabdf-cbda-4cf7-a518-13f619716d12 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 773.383386] env[62692]: DEBUG oslo_concurrency.lockutils [req-1a419b02-d71b-4d49-a6cf-73374a62d44d req-ea0f237f-1888-4aea-804c-e21bb6a3b97f service nova] Acquiring lock "d99737d8-2eb0-40ee-b61e-6c736c84ea59-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.383386] env[62692]: DEBUG oslo_concurrency.lockutils [req-1a419b02-d71b-4d49-a6cf-73374a62d44d req-ea0f237f-1888-4aea-804c-e21bb6a3b97f service nova] Lock "d99737d8-2eb0-40ee-b61e-6c736c84ea59-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.383386] env[62692]: DEBUG oslo_concurrency.lockutils [req-1a419b02-d71b-4d49-a6cf-73374a62d44d req-ea0f237f-1888-4aea-804c-e21bb6a3b97f service nova] Lock "d99737d8-2eb0-40ee-b61e-6c736c84ea59-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.383386] env[62692]: DEBUG nova.compute.manager [req-1a419b02-d71b-4d49-a6cf-73374a62d44d req-ea0f237f-1888-4aea-804c-e21bb6a3b97f service nova] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] No waiting events found dispatching network-vif-plugged-7f2cabdf-cbda-4cf7-a518-13f619716d12 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 773.383585] env[62692]: WARNING nova.compute.manager [req-1a419b02-d71b-4d49-a6cf-73374a62d44d req-ea0f237f-1888-4aea-804c-e21bb6a3b97f service nova] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Received unexpected event network-vif-plugged-7f2cabdf-cbda-4cf7-a518-13f619716d12 for instance with vm_state building and task_state spawning. [ 773.383725] env[62692]: DEBUG oslo_concurrency.lockutils [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.471s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.384292] env[62692]: DEBUG nova.compute.manager [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 773.393268] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.055s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.432089] env[62692]: DEBUG nova.network.neutron [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Successfully updated port: 7f2cabdf-cbda-4cf7-a518-13f619716d12 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 773.890438] env[62692]: DEBUG nova.compute.utils [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 773.891849] env[62692]: DEBUG nova.compute.manager [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 773.895298] env[62692]: DEBUG nova.network.neutron [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 773.935437] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "refresh_cache-d99737d8-2eb0-40ee-b61e-6c736c84ea59" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.935437] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired lock "refresh_cache-d99737d8-2eb0-40ee-b61e-6c736c84ea59" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.935644] env[62692]: DEBUG nova.network.neutron [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 773.998083] env[62692]: DEBUG nova.policy [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1399b3d529cb47458ebbf234bafe9a1f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9667b0b7abaa4bf0a7e54e9d540df728', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 774.305583] env[62692]: DEBUG nova.network.neutron [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Successfully created port: 30e027a7-86b8-4669-91fe-1dbd40ae54a5 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 774.322284] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83310b84-25fe-440f-9f02-525ed351656e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.329941] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03d94e7b-9f2c-40ae-a1fd-7563896061cf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.361280] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bdbb5ed-6ef9-41da-872e-74dfd3a04784 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.368858] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7876d7fe-14a1-47bb-9d0c-2e418752993c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.381747] env[62692]: DEBUG nova.compute.provider_tree [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 774.400857] env[62692]: DEBUG nova.compute.manager [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 774.494427] env[62692]: DEBUG nova.network.neutron [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 774.884338] env[62692]: DEBUG nova.scheduler.client.report [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 774.991420] env[62692]: DEBUG nova.network.neutron [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Updating instance_info_cache with network_info: [{"id": "7f2cabdf-cbda-4cf7-a518-13f619716d12", "address": "fa:16:3e:6b:22:0c", "network": {"id": "f70727d6-a9aa-44b0-a244-2533b3f96571", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-284105837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1037e5d436bd429391159b87a712f6e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f2cabdf-cb", "ovs_interfaceid": "7f2cabdf-cbda-4cf7-a518-13f619716d12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.389671] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.391452] env[62692]: ERROR nova.compute.manager [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 42704cff-466f-442d-8911-a6365882ee2d, please check neutron logs for more information. [ 775.391452] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] Traceback (most recent call last): [ 775.391452] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 775.391452] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] self.driver.spawn(context, instance, image_meta, [ 775.391452] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 775.391452] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 775.391452] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 775.391452] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] vm_ref = self.build_virtual_machine(instance, [ 775.391452] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 775.391452] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] vif_infos = vmwarevif.get_vif_info(self._session, [ 775.391452] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 775.391926] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] for vif in network_info: [ 775.391926] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 775.391926] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] return self._sync_wrapper(fn, *args, **kwargs) [ 775.391926] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 775.391926] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] self.wait() [ 775.391926] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 775.391926] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] self[:] = self._gt.wait() [ 775.391926] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 775.391926] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] return self._exit_event.wait() [ 775.391926] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 775.391926] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] current.throw(*self._exc) [ 775.391926] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 775.391926] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] result = function(*args, **kwargs) [ 775.392336] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 775.392336] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] return func(*args, **kwargs) [ 775.392336] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 775.392336] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] raise e [ 775.392336] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 775.392336] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] nwinfo = self.network_api.allocate_for_instance( [ 775.392336] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 775.392336] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] created_port_ids = self._update_ports_for_instance( [ 775.392336] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 775.392336] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] with excutils.save_and_reraise_exception(): [ 775.392336] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 775.392336] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] self.force_reraise() [ 775.392336] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 775.392727] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] raise self.value [ 775.392727] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 775.392727] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] updated_port = self._update_port( [ 775.392727] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 775.392727] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] _ensure_no_port_binding_failure(port) [ 775.392727] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 775.392727] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] raise exception.PortBindingFailed(port_id=port['id']) [ 775.392727] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] nova.exception.PortBindingFailed: Binding failed for port 42704cff-466f-442d-8911-a6365882ee2d, please check neutron logs for more information. [ 775.392727] env[62692]: ERROR nova.compute.manager [instance: 700d3723-f787-4770-a409-642cd8085e9c] [ 775.396017] env[62692]: DEBUG nova.compute.utils [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Binding failed for port 42704cff-466f-442d-8911-a6365882ee2d, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 775.396776] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.925s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.399159] env[62692]: INFO nova.compute.claims [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 775.402012] env[62692]: DEBUG nova.compute.manager [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Build of instance 700d3723-f787-4770-a409-642cd8085e9c was re-scheduled: Binding failed for port 42704cff-466f-442d-8911-a6365882ee2d, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 775.404103] env[62692]: DEBUG nova.compute.manager [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 775.404103] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Acquiring lock "refresh_cache-700d3723-f787-4770-a409-642cd8085e9c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.404103] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Acquired lock "refresh_cache-700d3723-f787-4770-a409-642cd8085e9c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.404103] env[62692]: DEBUG nova.network.neutron [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 775.409995] env[62692]: DEBUG nova.compute.manager [req-f5891b29-a8fd-4236-a8c3-a4f6220f443b req-f2a5accc-6808-438e-a43c-5f2d0fbe707b service nova] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Received event network-changed-7f2cabdf-cbda-4cf7-a518-13f619716d12 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 775.410213] env[62692]: DEBUG nova.compute.manager [req-f5891b29-a8fd-4236-a8c3-a4f6220f443b req-f2a5accc-6808-438e-a43c-5f2d0fbe707b service nova] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Refreshing instance network info cache due to event network-changed-7f2cabdf-cbda-4cf7-a518-13f619716d12. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 775.410464] env[62692]: DEBUG oslo_concurrency.lockutils [req-f5891b29-a8fd-4236-a8c3-a4f6220f443b req-f2a5accc-6808-438e-a43c-5f2d0fbe707b service nova] Acquiring lock "refresh_cache-d99737d8-2eb0-40ee-b61e-6c736c84ea59" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.411920] env[62692]: DEBUG nova.compute.manager [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 775.435736] env[62692]: DEBUG nova.virt.hardware [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 775.435982] env[62692]: DEBUG nova.virt.hardware [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 775.436152] env[62692]: DEBUG nova.virt.hardware [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 775.436341] env[62692]: DEBUG nova.virt.hardware [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 775.436486] env[62692]: DEBUG nova.virt.hardware [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 775.436634] env[62692]: DEBUG nova.virt.hardware [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 775.436842] env[62692]: DEBUG nova.virt.hardware [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 775.437010] env[62692]: DEBUG nova.virt.hardware [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 775.437191] env[62692]: DEBUG nova.virt.hardware [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 775.437357] env[62692]: DEBUG nova.virt.hardware [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 775.437530] env[62692]: DEBUG nova.virt.hardware [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 775.438467] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7b26b79-16e6-49b6-b85c-44b6f22f5c08 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.447745] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d170816-2e39-4141-bf1f-156330c45b1c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.493291] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Releasing lock "refresh_cache-d99737d8-2eb0-40ee-b61e-6c736c84ea59" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.493632] env[62692]: DEBUG nova.compute.manager [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Instance network_info: |[{"id": "7f2cabdf-cbda-4cf7-a518-13f619716d12", "address": "fa:16:3e:6b:22:0c", "network": {"id": "f70727d6-a9aa-44b0-a244-2533b3f96571", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-284105837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1037e5d436bd429391159b87a712f6e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f2cabdf-cb", "ovs_interfaceid": "7f2cabdf-cbda-4cf7-a518-13f619716d12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 775.493935] env[62692]: DEBUG oslo_concurrency.lockutils [req-f5891b29-a8fd-4236-a8c3-a4f6220f443b req-f2a5accc-6808-438e-a43c-5f2d0fbe707b service nova] Acquired lock "refresh_cache-d99737d8-2eb0-40ee-b61e-6c736c84ea59" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.494126] env[62692]: DEBUG nova.network.neutron [req-f5891b29-a8fd-4236-a8c3-a4f6220f443b req-f2a5accc-6808-438e-a43c-5f2d0fbe707b service nova] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Refreshing network info cache for port 7f2cabdf-cbda-4cf7-a518-13f619716d12 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 775.496076] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6b:22:0c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f44b2fa3-6730-4b87-8839-947eff21213f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7f2cabdf-cbda-4cf7-a518-13f619716d12', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 775.504741] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Creating folder: Project (1037e5d436bd429391159b87a712f6e1). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 775.509497] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dc8d317b-4053-48ec-b1e5-b8a5f7c5440d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.521863] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Created folder: Project (1037e5d436bd429391159b87a712f6e1) in parent group-v248868. [ 775.522088] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Creating folder: Instances. Parent ref: group-v248876. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 775.522327] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-16853ad2-fbbc-43b1-85e5-1d1a7e653af0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.530770] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Created folder: Instances in parent group-v248876. [ 775.530770] env[62692]: DEBUG oslo.service.loopingcall [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 775.530976] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 775.531207] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dc0cf999-fc8f-4c1b-a99e-525c7a386912 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.554623] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 775.554623] env[62692]: value = "task-1140899" [ 775.554623] env[62692]: _type = "Task" [ 775.554623] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.565431] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140899, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.758885] env[62692]: DEBUG nova.network.neutron [req-f5891b29-a8fd-4236-a8c3-a4f6220f443b req-f2a5accc-6808-438e-a43c-5f2d0fbe707b service nova] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Updated VIF entry in instance network info cache for port 7f2cabdf-cbda-4cf7-a518-13f619716d12. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 775.759849] env[62692]: DEBUG nova.network.neutron [req-f5891b29-a8fd-4236-a8c3-a4f6220f443b req-f2a5accc-6808-438e-a43c-5f2d0fbe707b service nova] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Updating instance_info_cache with network_info: [{"id": "7f2cabdf-cbda-4cf7-a518-13f619716d12", "address": "fa:16:3e:6b:22:0c", "network": {"id": "f70727d6-a9aa-44b0-a244-2533b3f96571", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-284105837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1037e5d436bd429391159b87a712f6e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f2cabdf-cb", "ovs_interfaceid": "7f2cabdf-cbda-4cf7-a518-13f619716d12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.922604] env[62692]: DEBUG nova.network.neutron [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 775.970117] env[62692]: DEBUG nova.network.neutron [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.067723] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140899, 'name': CreateVM_Task, 'duration_secs': 0.309537} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.067979] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 776.076516] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.076891] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.077105] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 776.077363] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f8299e1-9c4a-43e7-aaca-246603071bdb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.082663] env[62692]: DEBUG oslo_vmware.api [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 776.082663] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52c75417-644c-e7f2-756e-7407a52967f4" [ 776.082663] env[62692]: _type = "Task" [ 776.082663] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.091847] env[62692]: DEBUG oslo_vmware.api [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52c75417-644c-e7f2-756e-7407a52967f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.265239] env[62692]: DEBUG oslo_concurrency.lockutils [req-f5891b29-a8fd-4236-a8c3-a4f6220f443b req-f2a5accc-6808-438e-a43c-5f2d0fbe707b service nova] Releasing lock "refresh_cache-d99737d8-2eb0-40ee-b61e-6c736c84ea59" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.340261] env[62692]: DEBUG nova.network.neutron [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Successfully updated port: 30e027a7-86b8-4669-91fe-1dbd40ae54a5 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 776.472141] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Releasing lock "refresh_cache-700d3723-f787-4770-a409-642cd8085e9c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.472387] env[62692]: DEBUG nova.compute.manager [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 776.472556] env[62692]: DEBUG nova.compute.manager [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 776.472722] env[62692]: DEBUG nova.network.neutron [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 776.489134] env[62692]: DEBUG nova.network.neutron [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 776.592526] env[62692]: DEBUG oslo_vmware.api [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52c75417-644c-e7f2-756e-7407a52967f4, 'name': SearchDatastore_Task, 'duration_secs': 0.010443} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.594431] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.594669] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 776.594900] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.595055] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.595240] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 776.595655] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-60360089-f308-4e10-8098-821b88c05113 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.602919] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 776.603107] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 776.605716] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45760ea6-d342-4256-a0a7-00233014e262 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.610509] env[62692]: DEBUG oslo_vmware.api [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 776.610509] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52b32ff7-beab-d849-67f3-e0050b1b49dd" [ 776.610509] env[62692]: _type = "Task" [ 776.610509] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.617970] env[62692]: DEBUG oslo_vmware.api [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52b32ff7-beab-d849-67f3-e0050b1b49dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.710733] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3aa9804-4c7c-4eeb-9494-4b550d901c16 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.720207] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51b08aed-033d-44bb-a16b-ba022823b316 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.752888] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1963f49-07a5-409f-b524-8d7521537fe3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.760872] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a122b70-535a-4579-bf71-d83ca97f13b7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.774325] env[62692]: DEBUG nova.compute.provider_tree [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.843322] env[62692]: DEBUG oslo_concurrency.lockutils [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Acquiring lock "refresh_cache-628cc505-3edf-4066-91be-da009ebcf219" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.843369] env[62692]: DEBUG oslo_concurrency.lockutils [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Acquired lock "refresh_cache-628cc505-3edf-4066-91be-da009ebcf219" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.843518] env[62692]: DEBUG nova.network.neutron [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 776.991091] env[62692]: DEBUG nova.network.neutron [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.121518] env[62692]: DEBUG oslo_vmware.api [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52b32ff7-beab-d849-67f3-e0050b1b49dd, 'name': SearchDatastore_Task, 'duration_secs': 0.007583} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.121868] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ed90ac5-15c9-45d8-999b-ee95c7ea089f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.127360] env[62692]: DEBUG oslo_vmware.api [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 777.127360] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]522c0396-a5ed-0487-391a-196f6c71a681" [ 777.127360] env[62692]: _type = "Task" [ 777.127360] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.134872] env[62692]: DEBUG oslo_vmware.api [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]522c0396-a5ed-0487-391a-196f6c71a681, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.279807] env[62692]: DEBUG nova.scheduler.client.report [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 777.396349] env[62692]: DEBUG nova.network.neutron [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 777.437237] env[62692]: DEBUG nova.compute.manager [req-1e289d8c-6e84-45bb-a82f-c133033778b4 req-d767b4b9-61fc-47f9-ba69-968118e1eb6c service nova] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Received event network-vif-plugged-30e027a7-86b8-4669-91fe-1dbd40ae54a5 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 777.437465] env[62692]: DEBUG oslo_concurrency.lockutils [req-1e289d8c-6e84-45bb-a82f-c133033778b4 req-d767b4b9-61fc-47f9-ba69-968118e1eb6c service nova] Acquiring lock "628cc505-3edf-4066-91be-da009ebcf219-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.437691] env[62692]: DEBUG oslo_concurrency.lockutils [req-1e289d8c-6e84-45bb-a82f-c133033778b4 req-d767b4b9-61fc-47f9-ba69-968118e1eb6c service nova] Lock "628cc505-3edf-4066-91be-da009ebcf219-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.437852] env[62692]: DEBUG oslo_concurrency.lockutils [req-1e289d8c-6e84-45bb-a82f-c133033778b4 req-d767b4b9-61fc-47f9-ba69-968118e1eb6c service nova] Lock "628cc505-3edf-4066-91be-da009ebcf219-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.438024] env[62692]: DEBUG nova.compute.manager [req-1e289d8c-6e84-45bb-a82f-c133033778b4 req-d767b4b9-61fc-47f9-ba69-968118e1eb6c service nova] [instance: 628cc505-3edf-4066-91be-da009ebcf219] No waiting events found dispatching network-vif-plugged-30e027a7-86b8-4669-91fe-1dbd40ae54a5 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 777.438199] env[62692]: WARNING nova.compute.manager [req-1e289d8c-6e84-45bb-a82f-c133033778b4 req-d767b4b9-61fc-47f9-ba69-968118e1eb6c service nova] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Received unexpected event network-vif-plugged-30e027a7-86b8-4669-91fe-1dbd40ae54a5 for instance with vm_state building and task_state spawning. [ 777.438389] env[62692]: DEBUG nova.compute.manager [req-1e289d8c-6e84-45bb-a82f-c133033778b4 req-d767b4b9-61fc-47f9-ba69-968118e1eb6c service nova] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Received event network-changed-30e027a7-86b8-4669-91fe-1dbd40ae54a5 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 777.438537] env[62692]: DEBUG nova.compute.manager [req-1e289d8c-6e84-45bb-a82f-c133033778b4 req-d767b4b9-61fc-47f9-ba69-968118e1eb6c service nova] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Refreshing instance network info cache due to event network-changed-30e027a7-86b8-4669-91fe-1dbd40ae54a5. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 777.438704] env[62692]: DEBUG oslo_concurrency.lockutils [req-1e289d8c-6e84-45bb-a82f-c133033778b4 req-d767b4b9-61fc-47f9-ba69-968118e1eb6c service nova] Acquiring lock "refresh_cache-628cc505-3edf-4066-91be-da009ebcf219" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.493601] env[62692]: INFO nova.compute.manager [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] [instance: 700d3723-f787-4770-a409-642cd8085e9c] Took 1.02 seconds to deallocate network for instance. [ 777.570039] env[62692]: DEBUG nova.network.neutron [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Updating instance_info_cache with network_info: [{"id": "30e027a7-86b8-4669-91fe-1dbd40ae54a5", "address": "fa:16:3e:2e:69:5e", "network": {"id": "7e7a1dcd-36ba-4a25-affa-6bd7fe895ade", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-508011613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9667b0b7abaa4bf0a7e54e9d540df728", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30e027a7-86", "ovs_interfaceid": "30e027a7-86b8-4669-91fe-1dbd40ae54a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.637129] env[62692]: DEBUG oslo_vmware.api [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]522c0396-a5ed-0487-391a-196f6c71a681, 'name': SearchDatastore_Task, 'duration_secs': 0.008722} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.637403] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.637660] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] d99737d8-2eb0-40ee-b61e-6c736c84ea59/d99737d8-2eb0-40ee-b61e-6c736c84ea59.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 777.637903] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2f9c07b1-2194-4668-88bd-29a0661722d2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.645122] env[62692]: DEBUG oslo_vmware.api [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 777.645122] env[62692]: value = "task-1140900" [ 777.645122] env[62692]: _type = "Task" [ 777.645122] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.653066] env[62692]: DEBUG oslo_vmware.api [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1140900, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.785121] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.388s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.785708] env[62692]: DEBUG nova.compute.manager [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 777.788519] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.319s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.789990] env[62692]: INFO nova.compute.claims [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 778.073865] env[62692]: DEBUG oslo_concurrency.lockutils [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Releasing lock "refresh_cache-628cc505-3edf-4066-91be-da009ebcf219" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.073865] env[62692]: DEBUG nova.compute.manager [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Instance network_info: |[{"id": "30e027a7-86b8-4669-91fe-1dbd40ae54a5", "address": "fa:16:3e:2e:69:5e", "network": {"id": "7e7a1dcd-36ba-4a25-affa-6bd7fe895ade", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-508011613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9667b0b7abaa4bf0a7e54e9d540df728", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30e027a7-86", "ovs_interfaceid": "30e027a7-86b8-4669-91fe-1dbd40ae54a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 778.074063] env[62692]: DEBUG oslo_concurrency.lockutils [req-1e289d8c-6e84-45bb-a82f-c133033778b4 req-d767b4b9-61fc-47f9-ba69-968118e1eb6c service nova] Acquired lock "refresh_cache-628cc505-3edf-4066-91be-da009ebcf219" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.074107] env[62692]: DEBUG nova.network.neutron [req-1e289d8c-6e84-45bb-a82f-c133033778b4 req-d767b4b9-61fc-47f9-ba69-968118e1eb6c service nova] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Refreshing network info cache for port 30e027a7-86b8-4669-91fe-1dbd40ae54a5 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 778.075365] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2e:69:5e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4a8f5363-be3a-4f92-9ccf-33bb0c8113b3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '30e027a7-86b8-4669-91fe-1dbd40ae54a5', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 778.082949] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Creating folder: Project (9667b0b7abaa4bf0a7e54e9d540df728). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 778.086196] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a04e8004-b9a7-42e4-b9c4-a54efd482af0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.099039] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Created folder: Project (9667b0b7abaa4bf0a7e54e9d540df728) in parent group-v248868. [ 778.099235] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Creating folder: Instances. Parent ref: group-v248879. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 778.099440] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b466e4cc-6839-4ebe-98cd-7b6ec09d5f44 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.109324] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Created folder: Instances in parent group-v248879. [ 778.109547] env[62692]: DEBUG oslo.service.loopingcall [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 778.109737] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 778.109940] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3b025df5-a39e-4bdf-a215-3c7742523d09 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.132280] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 778.132280] env[62692]: value = "task-1140903" [ 778.132280] env[62692]: _type = "Task" [ 778.132280] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.140109] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140903, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.154106] env[62692]: DEBUG oslo_vmware.api [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1140900, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.449066} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.154285] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] d99737d8-2eb0-40ee-b61e-6c736c84ea59/d99737d8-2eb0-40ee-b61e-6c736c84ea59.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 778.154485] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 778.154735] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8f31e67c-0b8a-4e63-8d7b-6197eb002ef2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.161938] env[62692]: DEBUG oslo_vmware.api [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 778.161938] env[62692]: value = "task-1140904" [ 778.161938] env[62692]: _type = "Task" [ 778.161938] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.168919] env[62692]: DEBUG oslo_vmware.api [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1140904, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.294530] env[62692]: DEBUG nova.compute.utils [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 778.296014] env[62692]: DEBUG nova.compute.manager [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 778.296457] env[62692]: DEBUG nova.network.neutron [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 778.382088] env[62692]: DEBUG nova.policy [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '17517e7cf7174a7cb3fa45381c91f99f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fc06bb980c584d2980ee5eb48e678772', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 778.431927] env[62692]: DEBUG nova.network.neutron [req-1e289d8c-6e84-45bb-a82f-c133033778b4 req-d767b4b9-61fc-47f9-ba69-968118e1eb6c service nova] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Updated VIF entry in instance network info cache for port 30e027a7-86b8-4669-91fe-1dbd40ae54a5. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 778.432341] env[62692]: DEBUG nova.network.neutron [req-1e289d8c-6e84-45bb-a82f-c133033778b4 req-d767b4b9-61fc-47f9-ba69-968118e1eb6c service nova] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Updating instance_info_cache with network_info: [{"id": "30e027a7-86b8-4669-91fe-1dbd40ae54a5", "address": "fa:16:3e:2e:69:5e", "network": {"id": "7e7a1dcd-36ba-4a25-affa-6bd7fe895ade", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-508011613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9667b0b7abaa4bf0a7e54e9d540df728", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30e027a7-86", "ovs_interfaceid": "30e027a7-86b8-4669-91fe-1dbd40ae54a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.527048] env[62692]: INFO nova.scheduler.client.report [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Deleted allocations for instance 700d3723-f787-4770-a409-642cd8085e9c [ 778.641875] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140903, 'name': CreateVM_Task, 'duration_secs': 0.300978} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.642059] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 778.642742] env[62692]: DEBUG oslo_concurrency.lockutils [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.642901] env[62692]: DEBUG oslo_concurrency.lockutils [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.643247] env[62692]: DEBUG oslo_concurrency.lockutils [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 778.643542] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a88e7ccc-5054-4582-b388-f5c6e313fbf7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.648601] env[62692]: DEBUG oslo_vmware.api [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Waiting for the task: (returnval){ [ 778.648601] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52155eaf-8051-b7b7-2253-f3901e82f136" [ 778.648601] env[62692]: _type = "Task" [ 778.648601] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.657703] env[62692]: DEBUG oslo_vmware.api [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52155eaf-8051-b7b7-2253-f3901e82f136, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.669203] env[62692]: DEBUG oslo_vmware.api [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1140904, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066995} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.669484] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 778.670199] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cad943c8-189a-4425-aef4-895158b89f87 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.693284] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Reconfiguring VM instance instance-00000021 to attach disk [datastore2] d99737d8-2eb0-40ee-b61e-6c736c84ea59/d99737d8-2eb0-40ee-b61e-6c736c84ea59.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 778.694504] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cbe0c213-7192-4cb6-8254-8448412a5c3e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.713690] env[62692]: DEBUG oslo_vmware.api [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 778.713690] env[62692]: value = "task-1140905" [ 778.713690] env[62692]: _type = "Task" [ 778.713690] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.723311] env[62692]: DEBUG oslo_vmware.api [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1140905, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.783545] env[62692]: DEBUG nova.network.neutron [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Successfully created port: 354f1e83-aa2d-441f-8461-937a689e16c2 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 778.801644] env[62692]: DEBUG nova.compute.manager [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 778.936657] env[62692]: DEBUG oslo_concurrency.lockutils [req-1e289d8c-6e84-45bb-a82f-c133033778b4 req-d767b4b9-61fc-47f9-ba69-968118e1eb6c service nova] Releasing lock "refresh_cache-628cc505-3edf-4066-91be-da009ebcf219" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.037748] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ad555fb6-d839-4df7-a05a-e2400c56dea1 tempest-TenantUsagesTestJSON-697824277 tempest-TenantUsagesTestJSON-697824277-project-member] Lock "700d3723-f787-4770-a409-642cd8085e9c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.491s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.160898] env[62692]: DEBUG oslo_vmware.api [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52155eaf-8051-b7b7-2253-f3901e82f136, 'name': SearchDatastore_Task, 'duration_secs': 0.008539} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.161725] env[62692]: DEBUG oslo_concurrency.lockutils [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.161725] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 779.161854] env[62692]: DEBUG oslo_concurrency.lockutils [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.161914] env[62692]: DEBUG oslo_concurrency.lockutils [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.162388] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 779.162388] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ac1a33f9-a1fe-42d4-ad9b-f2b9ad2043cb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.165763] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e96544b-3249-432e-b94f-6a78d2e08382 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.169640] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 779.169816] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 779.171861] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2eb8b32e-95e6-46a9-b749-e1c3b95285a2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.177195] env[62692]: DEBUG oslo_vmware.api [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Waiting for the task: (returnval){ [ 779.177195] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5239d5b8-1b05-9bf1-461c-b480fa68ff3d" [ 779.177195] env[62692]: _type = "Task" [ 779.177195] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.182862] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9073fe2c-7355-4154-b324-6bbf7e484390 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.189278] env[62692]: DEBUG oslo_vmware.api [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5239d5b8-1b05-9bf1-461c-b480fa68ff3d, 'name': SearchDatastore_Task, 'duration_secs': 0.00784} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.190249] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd78a26f-f92f-4bc7-8d63-0d403f2c2d9d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.221276] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7dc28cd-68f2-44b0-b558-b07952e0b5b9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.225423] env[62692]: DEBUG oslo_vmware.api [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Waiting for the task: (returnval){ [ 779.225423] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]527c2f07-b648-50a7-d537-27c20d24f925" [ 779.225423] env[62692]: _type = "Task" [ 779.225423] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.233971] env[62692]: DEBUG oslo_vmware.api [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1140905, 'name': ReconfigVM_Task, 'duration_secs': 0.27807} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.234562] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Reconfigured VM instance instance-00000021 to attach disk [datastore2] d99737d8-2eb0-40ee-b61e-6c736c84ea59/d99737d8-2eb0-40ee-b61e-6c736c84ea59.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 779.235513] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b0d9b20d-b43c-4b45-9bfb-42be19719224 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.243919] env[62692]: DEBUG oslo_vmware.api [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]527c2f07-b648-50a7-d537-27c20d24f925, 'name': SearchDatastore_Task, 'duration_secs': 0.009918} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.245272] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dd568a0-f1b2-4ea5-9655-68a0a48cf14f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.249403] env[62692]: DEBUG oslo_concurrency.lockutils [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.249763] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 628cc505-3edf-4066-91be-da009ebcf219/628cc505-3edf-4066-91be-da009ebcf219.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 779.250827] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9a1f6891-f91e-4699-af08-b3cb7d516390 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.252618] env[62692]: DEBUG oslo_vmware.api [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 779.252618] env[62692]: value = "task-1140906" [ 779.252618] env[62692]: _type = "Task" [ 779.252618] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.263772] env[62692]: DEBUG nova.compute.provider_tree [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 779.266347] env[62692]: DEBUG oslo_vmware.api [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Waiting for the task: (returnval){ [ 779.266347] env[62692]: value = "task-1140907" [ 779.266347] env[62692]: _type = "Task" [ 779.266347] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.272554] env[62692]: DEBUG oslo_vmware.api [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1140906, 'name': Rename_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.278079] env[62692]: DEBUG oslo_vmware.api [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1140907, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.541626] env[62692]: DEBUG nova.compute.manager [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 779.762917] env[62692]: DEBUG oslo_vmware.api [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1140906, 'name': Rename_Task, 'duration_secs': 0.132955} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.763385] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 779.763739] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-100429ed-6aeb-4fb8-8de9-e7f2bf7df657 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.768469] env[62692]: DEBUG nova.scheduler.client.report [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 779.776114] env[62692]: DEBUG oslo_vmware.api [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 779.776114] env[62692]: value = "task-1140908" [ 779.776114] env[62692]: _type = "Task" [ 779.776114] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.784634] env[62692]: DEBUG oslo_vmware.api [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1140907, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.483828} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.785763] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 628cc505-3edf-4066-91be-da009ebcf219/628cc505-3edf-4066-91be-da009ebcf219.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 779.786080] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 779.786569] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-23879219-9f0d-440d-bb6f-3c347ba5c7ff {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.791995] env[62692]: DEBUG oslo_vmware.api [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1140908, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.797544] env[62692]: DEBUG oslo_vmware.api [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Waiting for the task: (returnval){ [ 779.797544] env[62692]: value = "task-1140909" [ 779.797544] env[62692]: _type = "Task" [ 779.797544] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.806822] env[62692]: DEBUG oslo_vmware.api [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1140909, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.812228] env[62692]: DEBUG nova.compute.manager [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 779.839699] env[62692]: DEBUG nova.virt.hardware [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 779.839840] env[62692]: DEBUG nova.virt.hardware [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 779.839995] env[62692]: DEBUG nova.virt.hardware [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 779.840251] env[62692]: DEBUG nova.virt.hardware [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 779.840403] env[62692]: DEBUG nova.virt.hardware [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 779.840592] env[62692]: DEBUG nova.virt.hardware [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 779.840837] env[62692]: DEBUG nova.virt.hardware [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 779.840998] env[62692]: DEBUG nova.virt.hardware [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 779.841187] env[62692]: DEBUG nova.virt.hardware [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 779.841351] env[62692]: DEBUG nova.virt.hardware [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 779.841565] env[62692]: DEBUG nova.virt.hardware [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 779.842460] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-087da7f8-5218-43d3-85d0-0f6225dc690c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.850726] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b5c176e-b299-45c7-9590-24a6b0363142 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.067282] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.262484] env[62692]: DEBUG nova.compute.manager [req-305fa76e-1318-484c-b71e-69dcf87594a7 req-10de9c58-6bad-46d3-922c-813be405e736 service nova] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Received event network-vif-plugged-354f1e83-aa2d-441f-8461-937a689e16c2 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 780.263978] env[62692]: DEBUG oslo_concurrency.lockutils [req-305fa76e-1318-484c-b71e-69dcf87594a7 req-10de9c58-6bad-46d3-922c-813be405e736 service nova] Acquiring lock "abee88ff-5cf5-4bf1-91e0-93b19cf30046-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.264569] env[62692]: DEBUG oslo_concurrency.lockutils [req-305fa76e-1318-484c-b71e-69dcf87594a7 req-10de9c58-6bad-46d3-922c-813be405e736 service nova] Lock "abee88ff-5cf5-4bf1-91e0-93b19cf30046-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.265235] env[62692]: DEBUG oslo_concurrency.lockutils [req-305fa76e-1318-484c-b71e-69dcf87594a7 req-10de9c58-6bad-46d3-922c-813be405e736 service nova] Lock "abee88ff-5cf5-4bf1-91e0-93b19cf30046-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.265537] env[62692]: DEBUG nova.compute.manager [req-305fa76e-1318-484c-b71e-69dcf87594a7 req-10de9c58-6bad-46d3-922c-813be405e736 service nova] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] No waiting events found dispatching network-vif-plugged-354f1e83-aa2d-441f-8461-937a689e16c2 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 780.265808] env[62692]: WARNING nova.compute.manager [req-305fa76e-1318-484c-b71e-69dcf87594a7 req-10de9c58-6bad-46d3-922c-813be405e736 service nova] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Received unexpected event network-vif-plugged-354f1e83-aa2d-441f-8461-937a689e16c2 for instance with vm_state building and task_state spawning. [ 780.281019] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.490s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.281019] env[62692]: DEBUG nova.compute.manager [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 780.281487] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.811s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.283386] env[62692]: INFO nova.compute.claims [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 780.297621] env[62692]: DEBUG oslo_vmware.api [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1140908, 'name': PowerOnVM_Task, 'duration_secs': 0.505311} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.298038] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 780.298247] env[62692]: INFO nova.compute.manager [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Took 7.36 seconds to spawn the instance on the hypervisor. [ 780.298618] env[62692]: DEBUG nova.compute.manager [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 780.299481] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02b259ed-2a2a-452e-bfe2-64a945f56b77 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.314614] env[62692]: DEBUG oslo_vmware.api [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1140909, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065839} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.315151] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 780.315890] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a5af5c1-9409-4bb5-8b1a-01256b9a5d9e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.339818] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Reconfiguring VM instance instance-00000022 to attach disk [datastore2] 628cc505-3edf-4066-91be-da009ebcf219/628cc505-3edf-4066-91be-da009ebcf219.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 780.341236] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be95713e-04fe-4541-a1af-67168a25bc98 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.361980] env[62692]: DEBUG oslo_vmware.api [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Waiting for the task: (returnval){ [ 780.361980] env[62692]: value = "task-1140910" [ 780.361980] env[62692]: _type = "Task" [ 780.361980] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.366087] env[62692]: DEBUG nova.network.neutron [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Successfully updated port: 354f1e83-aa2d-441f-8461-937a689e16c2 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 780.372779] env[62692]: DEBUG oslo_vmware.api [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1140910, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.795707] env[62692]: DEBUG nova.compute.utils [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 780.797223] env[62692]: DEBUG nova.compute.manager [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 780.797401] env[62692]: DEBUG nova.network.neutron [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 780.822389] env[62692]: INFO nova.compute.manager [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Took 41.24 seconds to build instance. [ 780.868602] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Acquiring lock "refresh_cache-abee88ff-5cf5-4bf1-91e0-93b19cf30046" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.868969] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Acquired lock "refresh_cache-abee88ff-5cf5-4bf1-91e0-93b19cf30046" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.869284] env[62692]: DEBUG nova.network.neutron [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 780.884901] env[62692]: DEBUG oslo_vmware.api [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1140910, 'name': ReconfigVM_Task, 'duration_secs': 0.274493} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.885713] env[62692]: DEBUG nova.policy [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b40c8976b544c059458aa4e0dbab160', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9cb8ae08824245a7911de7455e7fcb68', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 780.887106] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Reconfigured VM instance instance-00000022 to attach disk [datastore2] 628cc505-3edf-4066-91be-da009ebcf219/628cc505-3edf-4066-91be-da009ebcf219.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 780.888070] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1596bc59-e827-406b-b431-0ac175f3bace {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.896387] env[62692]: DEBUG oslo_vmware.api [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Waiting for the task: (returnval){ [ 780.896387] env[62692]: value = "task-1140911" [ 780.896387] env[62692]: _type = "Task" [ 780.896387] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.903196] env[62692]: DEBUG oslo_vmware.api [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1140911, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.952721] env[62692]: DEBUG nova.network.neutron [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 781.238639] env[62692]: DEBUG nova.network.neutron [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Updating instance_info_cache with network_info: [{"id": "354f1e83-aa2d-441f-8461-937a689e16c2", "address": "fa:16:3e:7a:37:15", "network": {"id": "14d97ab8-1b79-4ef2-a354-6398d00bdeba", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-707295669-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fc06bb980c584d2980ee5eb48e678772", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap354f1e83-aa", "ovs_interfaceid": "354f1e83-aa2d-441f-8461-937a689e16c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.302880] env[62692]: DEBUG nova.compute.manager [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 781.324591] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5ee6b040-e463-4319-920f-50ea531417e9 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "d99737d8-2eb0-40ee-b61e-6c736c84ea59" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.798s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.377138] env[62692]: DEBUG nova.network.neutron [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Successfully created port: f19989c4-0ce9-47ae-989b-0faab24bc9b1 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 781.406512] env[62692]: DEBUG oslo_vmware.api [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1140911, 'name': Rename_Task, 'duration_secs': 0.135423} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.406991] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 781.409217] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-12976b4b-9ea0-479f-a934-551712f87cd8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.416575] env[62692]: DEBUG oslo_vmware.api [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Waiting for the task: (returnval){ [ 781.416575] env[62692]: value = "task-1140912" [ 781.416575] env[62692]: _type = "Task" [ 781.416575] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.425411] env[62692]: DEBUG oslo_vmware.api [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1140912, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.716677] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba925410-8427-4458-bc57-4bf3b1831676 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.724561] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79325f28-e93b-48c6-80a1-ead14f61bb1e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.767576] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Releasing lock "refresh_cache-abee88ff-5cf5-4bf1-91e0-93b19cf30046" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.767989] env[62692]: DEBUG nova.compute.manager [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Instance network_info: |[{"id": "354f1e83-aa2d-441f-8461-937a689e16c2", "address": "fa:16:3e:7a:37:15", "network": {"id": "14d97ab8-1b79-4ef2-a354-6398d00bdeba", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-707295669-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fc06bb980c584d2980ee5eb48e678772", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap354f1e83-aa", "ovs_interfaceid": "354f1e83-aa2d-441f-8461-937a689e16c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 781.768806] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7a:37:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60bdba1a-14cf-46b2-9d8b-aeaf4d80c815', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '354f1e83-aa2d-441f-8461-937a689e16c2', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 781.776274] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Creating folder: Project (fc06bb980c584d2980ee5eb48e678772). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 781.777115] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e766b5b6-799a-4249-8642-211a59f385f7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.779926] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-36386faa-bd65-4612-b6b0-ddcc3beddc94 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.787901] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85ed5f0c-6230-4931-b6aa-a9fb81e75a98 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.793267] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Created folder: Project (fc06bb980c584d2980ee5eb48e678772) in parent group-v248868. [ 781.793458] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Creating folder: Instances. Parent ref: group-v248882. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 781.793987] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-daa4d714-1f15-4c1c-a07d-d30257d5e0f4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.805851] env[62692]: DEBUG nova.compute.provider_tree [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 781.816029] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Created folder: Instances in parent group-v248882. [ 781.816029] env[62692]: DEBUG oslo.service.loopingcall [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 781.816029] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 781.816029] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-780e82f5-c31a-4546-9a5b-9f36c484e725 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.829857] env[62692]: DEBUG nova.compute.manager [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 781.838184] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 781.838184] env[62692]: value = "task-1140915" [ 781.838184] env[62692]: _type = "Task" [ 781.838184] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.846782] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140915, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.927856] env[62692]: DEBUG oslo_vmware.api [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1140912, 'name': PowerOnVM_Task, 'duration_secs': 0.44871} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.928160] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 781.928369] env[62692]: INFO nova.compute.manager [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Took 6.52 seconds to spawn the instance on the hypervisor. [ 781.928582] env[62692]: DEBUG nova.compute.manager [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 781.929387] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f9f786-afee-4a5e-84b1-6370caa0d1a2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.288967] env[62692]: DEBUG nova.compute.manager [req-2df57831-c7f2-4ac9-b0a9-623e6a63a308 req-edc13c9a-971e-4525-9589-1cd4f9880dab service nova] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Received event network-changed-354f1e83-aa2d-441f-8461-937a689e16c2 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 782.289332] env[62692]: DEBUG nova.compute.manager [req-2df57831-c7f2-4ac9-b0a9-623e6a63a308 req-edc13c9a-971e-4525-9589-1cd4f9880dab service nova] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Refreshing instance network info cache due to event network-changed-354f1e83-aa2d-441f-8461-937a689e16c2. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 782.289785] env[62692]: DEBUG oslo_concurrency.lockutils [req-2df57831-c7f2-4ac9-b0a9-623e6a63a308 req-edc13c9a-971e-4525-9589-1cd4f9880dab service nova] Acquiring lock "refresh_cache-abee88ff-5cf5-4bf1-91e0-93b19cf30046" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.290095] env[62692]: DEBUG oslo_concurrency.lockutils [req-2df57831-c7f2-4ac9-b0a9-623e6a63a308 req-edc13c9a-971e-4525-9589-1cd4f9880dab service nova] Acquired lock "refresh_cache-abee88ff-5cf5-4bf1-91e0-93b19cf30046" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.290358] env[62692]: DEBUG nova.network.neutron [req-2df57831-c7f2-4ac9-b0a9-623e6a63a308 req-edc13c9a-971e-4525-9589-1cd4f9880dab service nova] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Refreshing network info cache for port 354f1e83-aa2d-441f-8461-937a689e16c2 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 782.316037] env[62692]: DEBUG nova.compute.manager [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 782.317283] env[62692]: DEBUG nova.scheduler.client.report [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 782.349438] env[62692]: DEBUG nova.virt.hardware [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 782.349438] env[62692]: DEBUG nova.virt.hardware [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 782.349438] env[62692]: DEBUG nova.virt.hardware [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 782.349813] env[62692]: DEBUG nova.virt.hardware [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 782.349813] env[62692]: DEBUG nova.virt.hardware [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 782.349813] env[62692]: DEBUG nova.virt.hardware [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 782.349813] env[62692]: DEBUG nova.virt.hardware [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 782.349813] env[62692]: DEBUG nova.virt.hardware [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 782.349988] env[62692]: DEBUG nova.virt.hardware [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 782.349988] env[62692]: DEBUG nova.virt.hardware [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 782.349988] env[62692]: DEBUG nova.virt.hardware [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 782.350580] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75242512-996f-4176-b701-e646a7720630 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.353808] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.360306] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140915, 'name': CreateVM_Task, 'duration_secs': 0.30514} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.362431] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 782.363730] env[62692]: DEBUG oslo_vmware.service [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-817f4876-c234-404f-b870-450c2cce93fa {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.367455] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc291af2-27f3-4654-9008-f0ee24e83e9b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.375312] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.375666] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.377029] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 782.377029] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c04f564d-bb57-4804-a752-a35edf524171 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.390453] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 782.390453] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52db3624-fecb-3137-7657-c8f730db4e78" [ 782.390453] env[62692]: _type = "Task" [ 782.390453] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.399939] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52db3624-fecb-3137-7657-c8f730db4e78, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.449931] env[62692]: INFO nova.compute.manager [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Took 40.48 seconds to build instance. [ 782.687508] env[62692]: DEBUG nova.compute.manager [req-427e6de5-b250-4287-8d83-f66cae270940 req-705f5e4f-87ff-42e0-b51b-25e53a800c84 service nova] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Received event network-changed-7f2cabdf-cbda-4cf7-a518-13f619716d12 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 782.687508] env[62692]: DEBUG nova.compute.manager [req-427e6de5-b250-4287-8d83-f66cae270940 req-705f5e4f-87ff-42e0-b51b-25e53a800c84 service nova] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Refreshing instance network info cache due to event network-changed-7f2cabdf-cbda-4cf7-a518-13f619716d12. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 782.687508] env[62692]: DEBUG oslo_concurrency.lockutils [req-427e6de5-b250-4287-8d83-f66cae270940 req-705f5e4f-87ff-42e0-b51b-25e53a800c84 service nova] Acquiring lock "refresh_cache-d99737d8-2eb0-40ee-b61e-6c736c84ea59" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.687508] env[62692]: DEBUG oslo_concurrency.lockutils [req-427e6de5-b250-4287-8d83-f66cae270940 req-705f5e4f-87ff-42e0-b51b-25e53a800c84 service nova] Acquired lock "refresh_cache-d99737d8-2eb0-40ee-b61e-6c736c84ea59" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.687508] env[62692]: DEBUG nova.network.neutron [req-427e6de5-b250-4287-8d83-f66cae270940 req-705f5e4f-87ff-42e0-b51b-25e53a800c84 service nova] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Refreshing network info cache for port 7f2cabdf-cbda-4cf7-a518-13f619716d12 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 782.823107] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.541s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.823714] env[62692]: DEBUG nova.compute.manager [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 782.827129] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.414s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.828922] env[62692]: INFO nova.compute.claims [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 782.902318] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.902668] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 782.905324] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.905324] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.905324] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 782.905324] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3c24a5b7-1700-4b84-8e37-20e158143363 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.912636] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 782.913714] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 782.916211] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae18b5e4-7a7b-44f6-babb-c864d87c274e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.923637] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a97bafa-5701-486f-a5e6-b61cfa3463de {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.928873] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 782.928873] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5212ad68-d66f-93fa-d746-b0808a316b44" [ 782.928873] env[62692]: _type = "Task" [ 782.928873] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.937736] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5212ad68-d66f-93fa-d746-b0808a316b44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.954582] env[62692]: DEBUG oslo_concurrency.lockutils [None req-870e9f3b-1cb8-4b43-b2b4-796bf9408a8e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Lock "628cc505-3edf-4066-91be-da009ebcf219" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.603s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.182305] env[62692]: DEBUG nova.network.neutron [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Successfully updated port: f19989c4-0ce9-47ae-989b-0faab24bc9b1 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 783.262315] env[62692]: DEBUG nova.network.neutron [req-2df57831-c7f2-4ac9-b0a9-623e6a63a308 req-edc13c9a-971e-4525-9589-1cd4f9880dab service nova] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Updated VIF entry in instance network info cache for port 354f1e83-aa2d-441f-8461-937a689e16c2. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 783.262315] env[62692]: DEBUG nova.network.neutron [req-2df57831-c7f2-4ac9-b0a9-623e6a63a308 req-edc13c9a-971e-4525-9589-1cd4f9880dab service nova] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Updating instance_info_cache with network_info: [{"id": "354f1e83-aa2d-441f-8461-937a689e16c2", "address": "fa:16:3e:7a:37:15", "network": {"id": "14d97ab8-1b79-4ef2-a354-6398d00bdeba", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-707295669-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fc06bb980c584d2980ee5eb48e678772", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap354f1e83-aa", "ovs_interfaceid": "354f1e83-aa2d-441f-8461-937a689e16c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.332842] env[62692]: DEBUG nova.compute.utils [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 783.336341] env[62692]: DEBUG nova.compute.manager [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Not allocating networking since 'none' was specified. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 783.439247] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Preparing fetch location {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 783.439562] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Creating directory with path [datastore1] vmware_temp/9657360f-fe08-490e-bf57-166bd1412e9f/81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 783.439728] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eb8679f3-091b-49d3-8b6b-8fd2f73a7a28 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.461712] env[62692]: DEBUG nova.compute.manager [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 783.474831] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Created directory with path [datastore1] vmware_temp/9657360f-fe08-490e-bf57-166bd1412e9f/81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 783.475074] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Fetch image to [datastore1] vmware_temp/9657360f-fe08-490e-bf57-166bd1412e9f/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/tmp-sparse.vmdk {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 783.475208] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Downloading image file data 81eb64fb-5a0c-47b7-9948-7dcb01b911bd to [datastore1] vmware_temp/9657360f-fe08-490e-bf57-166bd1412e9f/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/tmp-sparse.vmdk on the data store datastore1 {{(pid=62692) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 783.476009] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dee7f3c5-ffac-43cd-af88-cf1d52405550 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.483755] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10011001-f043-4f29-8f9e-3e2010d13945 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.497916] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74c43d68-119b-48ce-9627-bb03ff802d7c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.533828] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85927b3d-856c-449c-b171-203f1158ee30 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.540532] env[62692]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-8896d86b-64ba-43b7-b73b-20253a1a2ac9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.632493] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Downloading image file data 81eb64fb-5a0c-47b7-9948-7dcb01b911bd to the data store datastore1 {{(pid=62692) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 783.639651] env[62692]: DEBUG nova.network.neutron [req-427e6de5-b250-4287-8d83-f66cae270940 req-705f5e4f-87ff-42e0-b51b-25e53a800c84 service nova] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Updated VIF entry in instance network info cache for port 7f2cabdf-cbda-4cf7-a518-13f619716d12. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 783.640040] env[62692]: DEBUG nova.network.neutron [req-427e6de5-b250-4287-8d83-f66cae270940 req-705f5e4f-87ff-42e0-b51b-25e53a800c84 service nova] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Updating instance_info_cache with network_info: [{"id": "7f2cabdf-cbda-4cf7-a518-13f619716d12", "address": "fa:16:3e:6b:22:0c", "network": {"id": "f70727d6-a9aa-44b0-a244-2533b3f96571", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-284105837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1037e5d436bd429391159b87a712f6e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f2cabdf-cb", "ovs_interfaceid": "7f2cabdf-cbda-4cf7-a518-13f619716d12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.685678] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "refresh_cache-fa7bb883-0065-4ecc-9eb5-8c5443344a2a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.687034] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired lock "refresh_cache-fa7bb883-0065-4ecc-9eb5-8c5443344a2a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.687034] env[62692]: DEBUG nova.network.neutron [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 783.713055] env[62692]: DEBUG oslo_vmware.rw_handles [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/9657360f-fe08-490e-bf57-166bd1412e9f/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62692) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 783.777126] env[62692]: DEBUG oslo_concurrency.lockutils [req-2df57831-c7f2-4ac9-b0a9-623e6a63a308 req-edc13c9a-971e-4525-9589-1cd4f9880dab service nova] Releasing lock "refresh_cache-abee88ff-5cf5-4bf1-91e0-93b19cf30046" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.837052] env[62692]: DEBUG nova.compute.manager [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 783.983980] env[62692]: DEBUG oslo_concurrency.lockutils [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.143867] env[62692]: DEBUG oslo_concurrency.lockutils [req-427e6de5-b250-4287-8d83-f66cae270940 req-705f5e4f-87ff-42e0-b51b-25e53a800c84 service nova] Releasing lock "refresh_cache-d99737d8-2eb0-40ee-b61e-6c736c84ea59" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.225732] env[62692]: DEBUG nova.network.neutron [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 784.291051] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf20dbdd-ba94-4549-b140-dedb8c0f2584 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.304736] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3a51d31-3f33-4ec7-af76-becd07f479a9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.345112] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7bf6205-db0d-4354-9c8c-818e1a65c0eb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.353474] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c580f8cf-2a69-40b0-84da-d1adecd230c0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.372066] env[62692]: DEBUG nova.compute.provider_tree [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 784.380072] env[62692]: DEBUG oslo_vmware.rw_handles [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Completed reading data from the image iterator. {{(pid=62692) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 784.380072] env[62692]: DEBUG oslo_vmware.rw_handles [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Closing write handle for https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/9657360f-fe08-490e-bf57-166bd1412e9f/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62692) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 784.504827] env[62692]: DEBUG nova.compute.manager [req-0fed755b-3004-41dc-8695-a8b43b602f0c req-6df88dc2-1371-4311-a703-54beb3a717bb service nova] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Received event network-vif-plugged-f19989c4-0ce9-47ae-989b-0faab24bc9b1 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 784.505140] env[62692]: DEBUG oslo_concurrency.lockutils [req-0fed755b-3004-41dc-8695-a8b43b602f0c req-6df88dc2-1371-4311-a703-54beb3a717bb service nova] Acquiring lock "fa7bb883-0065-4ecc-9eb5-8c5443344a2a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.505483] env[62692]: DEBUG oslo_concurrency.lockutils [req-0fed755b-3004-41dc-8695-a8b43b602f0c req-6df88dc2-1371-4311-a703-54beb3a717bb service nova] Lock "fa7bb883-0065-4ecc-9eb5-8c5443344a2a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.506221] env[62692]: DEBUG oslo_concurrency.lockutils [req-0fed755b-3004-41dc-8695-a8b43b602f0c req-6df88dc2-1371-4311-a703-54beb3a717bb service nova] Lock "fa7bb883-0065-4ecc-9eb5-8c5443344a2a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.506221] env[62692]: DEBUG nova.compute.manager [req-0fed755b-3004-41dc-8695-a8b43b602f0c req-6df88dc2-1371-4311-a703-54beb3a717bb service nova] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] No waiting events found dispatching network-vif-plugged-f19989c4-0ce9-47ae-989b-0faab24bc9b1 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 784.506221] env[62692]: WARNING nova.compute.manager [req-0fed755b-3004-41dc-8695-a8b43b602f0c req-6df88dc2-1371-4311-a703-54beb3a717bb service nova] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Received unexpected event network-vif-plugged-f19989c4-0ce9-47ae-989b-0faab24bc9b1 for instance with vm_state building and task_state spawning. [ 784.508109] env[62692]: DEBUG nova.compute.manager [req-0fed755b-3004-41dc-8695-a8b43b602f0c req-6df88dc2-1371-4311-a703-54beb3a717bb service nova] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Received event network-changed-f19989c4-0ce9-47ae-989b-0faab24bc9b1 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 784.508109] env[62692]: DEBUG nova.compute.manager [req-0fed755b-3004-41dc-8695-a8b43b602f0c req-6df88dc2-1371-4311-a703-54beb3a717bb service nova] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Refreshing instance network info cache due to event network-changed-f19989c4-0ce9-47ae-989b-0faab24bc9b1. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 784.508109] env[62692]: DEBUG oslo_concurrency.lockutils [req-0fed755b-3004-41dc-8695-a8b43b602f0c req-6df88dc2-1371-4311-a703-54beb3a717bb service nova] Acquiring lock "refresh_cache-fa7bb883-0065-4ecc-9eb5-8c5443344a2a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.534025] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Downloaded image file data 81eb64fb-5a0c-47b7-9948-7dcb01b911bd to vmware_temp/9657360f-fe08-490e-bf57-166bd1412e9f/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/tmp-sparse.vmdk on the data store datastore1 {{(pid=62692) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 784.536636] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Caching image {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 784.536636] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Copying Virtual Disk [datastore1] vmware_temp/9657360f-fe08-490e-bf57-166bd1412e9f/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/tmp-sparse.vmdk to [datastore1] vmware_temp/9657360f-fe08-490e-bf57-166bd1412e9f/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 784.536636] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cea337e2-0310-4b1c-8bff-dcb932d9d161 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.544571] env[62692]: DEBUG nova.network.neutron [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Updating instance_info_cache with network_info: [{"id": "f19989c4-0ce9-47ae-989b-0faab24bc9b1", "address": "fa:16:3e:5a:54:50", "network": {"id": "f6352999-f943-431d-b54a-c9dc1d810a85", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-121037111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9cb8ae08824245a7911de7455e7fcb68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c842425c-544e-4ce2-9657-512723bd318e", "external-id": "nsx-vlan-transportzone-80", "segmentation_id": 80, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf19989c4-0c", "ovs_interfaceid": "f19989c4-0ce9-47ae-989b-0faab24bc9b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.547473] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 784.547473] env[62692]: value = "task-1140916" [ 784.547473] env[62692]: _type = "Task" [ 784.547473] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.558115] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140916, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.805539] env[62692]: DEBUG nova.compute.manager [req-35c39032-75c5-4879-ac62-ea84968a9957 req-e1848722-46d5-4c73-b588-aedd57fcfaa2 service nova] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Received event network-changed-30e027a7-86b8-4669-91fe-1dbd40ae54a5 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 784.805769] env[62692]: DEBUG nova.compute.manager [req-35c39032-75c5-4879-ac62-ea84968a9957 req-e1848722-46d5-4c73-b588-aedd57fcfaa2 service nova] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Refreshing instance network info cache due to event network-changed-30e027a7-86b8-4669-91fe-1dbd40ae54a5. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 784.805982] env[62692]: DEBUG oslo_concurrency.lockutils [req-35c39032-75c5-4879-ac62-ea84968a9957 req-e1848722-46d5-4c73-b588-aedd57fcfaa2 service nova] Acquiring lock "refresh_cache-628cc505-3edf-4066-91be-da009ebcf219" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.806596] env[62692]: DEBUG oslo_concurrency.lockutils [req-35c39032-75c5-4879-ac62-ea84968a9957 req-e1848722-46d5-4c73-b588-aedd57fcfaa2 service nova] Acquired lock "refresh_cache-628cc505-3edf-4066-91be-da009ebcf219" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.806786] env[62692]: DEBUG nova.network.neutron [req-35c39032-75c5-4879-ac62-ea84968a9957 req-e1848722-46d5-4c73-b588-aedd57fcfaa2 service nova] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Refreshing network info cache for port 30e027a7-86b8-4669-91fe-1dbd40ae54a5 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 784.863380] env[62692]: DEBUG nova.compute.manager [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 784.882031] env[62692]: DEBUG nova.scheduler.client.report [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 784.899594] env[62692]: DEBUG nova.virt.hardware [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 784.899920] env[62692]: DEBUG nova.virt.hardware [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 784.900093] env[62692]: DEBUG nova.virt.hardware [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 784.900294] env[62692]: DEBUG nova.virt.hardware [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 784.900414] env[62692]: DEBUG nova.virt.hardware [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 784.900562] env[62692]: DEBUG nova.virt.hardware [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 784.900810] env[62692]: DEBUG nova.virt.hardware [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 784.900968] env[62692]: DEBUG nova.virt.hardware [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 784.901282] env[62692]: DEBUG nova.virt.hardware [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 784.901496] env[62692]: DEBUG nova.virt.hardware [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 784.901724] env[62692]: DEBUG nova.virt.hardware [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 784.902618] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d9b40f-5f4e-4fd8-b1c2-680480645970 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.913951] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc674ebf-9e28-4fff-a948-07ac8655bd6a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.932451] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Instance VIF info [] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 784.939640] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Creating folder: Project (051e05abbe4049529b3dd72c8aec96ac). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 784.941043] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4af793d3-a21e-4dc2-a673-c8400bc44ee2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.952932] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Created folder: Project (051e05abbe4049529b3dd72c8aec96ac) in parent group-v248868. [ 784.952976] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Creating folder: Instances. Parent ref: group-v248885. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 784.953356] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d5f056b2-38e1-4e3d-a3ef-2b2c584dabad {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.963101] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Created folder: Instances in parent group-v248885. [ 784.963668] env[62692]: DEBUG oslo.service.loopingcall [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 784.963882] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 784.964158] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-383aaa9d-96f2-4bd9-b6fb-514675396a12 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.984020] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 784.984020] env[62692]: value = "task-1140919" [ 784.984020] env[62692]: _type = "Task" [ 784.984020] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.992273] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140919, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.048615] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Releasing lock "refresh_cache-fa7bb883-0065-4ecc-9eb5-8c5443344a2a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.048988] env[62692]: DEBUG nova.compute.manager [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Instance network_info: |[{"id": "f19989c4-0ce9-47ae-989b-0faab24bc9b1", "address": "fa:16:3e:5a:54:50", "network": {"id": "f6352999-f943-431d-b54a-c9dc1d810a85", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-121037111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9cb8ae08824245a7911de7455e7fcb68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c842425c-544e-4ce2-9657-512723bd318e", "external-id": "nsx-vlan-transportzone-80", "segmentation_id": 80, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf19989c4-0c", "ovs_interfaceid": "f19989c4-0ce9-47ae-989b-0faab24bc9b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 785.049449] env[62692]: DEBUG oslo_concurrency.lockutils [req-0fed755b-3004-41dc-8695-a8b43b602f0c req-6df88dc2-1371-4311-a703-54beb3a717bb service nova] Acquired lock "refresh_cache-fa7bb883-0065-4ecc-9eb5-8c5443344a2a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.049668] env[62692]: DEBUG nova.network.neutron [req-0fed755b-3004-41dc-8695-a8b43b602f0c req-6df88dc2-1371-4311-a703-54beb3a717bb service nova] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Refreshing network info cache for port f19989c4-0ce9-47ae-989b-0faab24bc9b1 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 785.050907] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5a:54:50', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c842425c-544e-4ce2-9657-512723bd318e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f19989c4-0ce9-47ae-989b-0faab24bc9b1', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 785.059412] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Creating folder: Project (9cb8ae08824245a7911de7455e7fcb68). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 785.066858] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0215856f-bb26-46fd-8a86-cc258f69270b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.074858] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140916, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.076706] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Created folder: Project (9cb8ae08824245a7911de7455e7fcb68) in parent group-v248868. [ 785.077121] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Creating folder: Instances. Parent ref: group-v248888. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 785.077383] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e2ecc335-193f-4679-a054-52824679b95d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.085999] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Created folder: Instances in parent group-v248888. [ 785.086395] env[62692]: DEBUG oslo.service.loopingcall [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 785.086591] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 785.086827] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-de4f3996-35d4-4c54-9e8e-81f2decb0c2d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.109081] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 785.109081] env[62692]: value = "task-1140922" [ 785.109081] env[62692]: _type = "Task" [ 785.109081] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.120855] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140922, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.389859] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.563s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.390373] env[62692]: DEBUG nova.compute.manager [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 785.398678] env[62692]: DEBUG oslo_concurrency.lockutils [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.837s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.429440] env[62692]: DEBUG nova.network.neutron [req-0fed755b-3004-41dc-8695-a8b43b602f0c req-6df88dc2-1371-4311-a703-54beb3a717bb service nova] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Updated VIF entry in instance network info cache for port f19989c4-0ce9-47ae-989b-0faab24bc9b1. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 785.429440] env[62692]: DEBUG nova.network.neutron [req-0fed755b-3004-41dc-8695-a8b43b602f0c req-6df88dc2-1371-4311-a703-54beb3a717bb service nova] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Updating instance_info_cache with network_info: [{"id": "f19989c4-0ce9-47ae-989b-0faab24bc9b1", "address": "fa:16:3e:5a:54:50", "network": {"id": "f6352999-f943-431d-b54a-c9dc1d810a85", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-121037111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9cb8ae08824245a7911de7455e7fcb68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c842425c-544e-4ce2-9657-512723bd318e", "external-id": "nsx-vlan-transportzone-80", "segmentation_id": 80, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf19989c4-0c", "ovs_interfaceid": "f19989c4-0ce9-47ae-989b-0faab24bc9b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.501235] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140919, 'name': CreateVM_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.560745] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140916, 'name': CopyVirtualDisk_Task} progress is 70%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.618883] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140922, 'name': CreateVM_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.690448] env[62692]: DEBUG nova.network.neutron [req-35c39032-75c5-4879-ac62-ea84968a9957 req-e1848722-46d5-4c73-b588-aedd57fcfaa2 service nova] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Updated VIF entry in instance network info cache for port 30e027a7-86b8-4669-91fe-1dbd40ae54a5. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 785.690838] env[62692]: DEBUG nova.network.neutron [req-35c39032-75c5-4879-ac62-ea84968a9957 req-e1848722-46d5-4c73-b588-aedd57fcfaa2 service nova] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Updating instance_info_cache with network_info: [{"id": "30e027a7-86b8-4669-91fe-1dbd40ae54a5", "address": "fa:16:3e:2e:69:5e", "network": {"id": "7e7a1dcd-36ba-4a25-affa-6bd7fe895ade", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-508011613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9667b0b7abaa4bf0a7e54e9d540df728", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30e027a7-86", "ovs_interfaceid": "30e027a7-86b8-4669-91fe-1dbd40ae54a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.900454] env[62692]: DEBUG nova.compute.utils [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 785.901976] env[62692]: DEBUG nova.compute.manager [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 785.902081] env[62692]: DEBUG nova.network.neutron [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 785.935476] env[62692]: DEBUG oslo_concurrency.lockutils [req-0fed755b-3004-41dc-8695-a8b43b602f0c req-6df88dc2-1371-4311-a703-54beb3a717bb service nova] Releasing lock "refresh_cache-fa7bb883-0065-4ecc-9eb5-8c5443344a2a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.935840] env[62692]: DEBUG nova.compute.manager [req-0fed755b-3004-41dc-8695-a8b43b602f0c req-6df88dc2-1371-4311-a703-54beb3a717bb service nova] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Received event network-changed-30e027a7-86b8-4669-91fe-1dbd40ae54a5 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 785.936063] env[62692]: DEBUG nova.compute.manager [req-0fed755b-3004-41dc-8695-a8b43b602f0c req-6df88dc2-1371-4311-a703-54beb3a717bb service nova] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Refreshing instance network info cache due to event network-changed-30e027a7-86b8-4669-91fe-1dbd40ae54a5. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 785.936300] env[62692]: DEBUG oslo_concurrency.lockutils [req-0fed755b-3004-41dc-8695-a8b43b602f0c req-6df88dc2-1371-4311-a703-54beb3a717bb service nova] Acquiring lock "refresh_cache-628cc505-3edf-4066-91be-da009ebcf219" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.972611] env[62692]: DEBUG nova.policy [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9e9f256c1ebf4b1ba966edb7b1df5f4f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '266c39fe585d41079d0278ecd1d1c6a4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 786.003395] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140919, 'name': CreateVM_Task, 'duration_secs': 0.786221} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.004189] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 786.004189] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.004417] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.004734] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 786.010019] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45576e99-72ae-430b-b7af-0b876e7090cb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.013223] env[62692]: DEBUG oslo_vmware.api [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Waiting for the task: (returnval){ [ 786.013223] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52cda0d7-0a28-5fb5-fbc1-1d420d4bba7c" [ 786.013223] env[62692]: _type = "Task" [ 786.013223] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.022669] env[62692]: DEBUG oslo_vmware.api [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52cda0d7-0a28-5fb5-fbc1-1d420d4bba7c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.066275] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140916, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.121610] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140922, 'name': CreateVM_Task, 'duration_secs': 0.688774} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.121610] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 786.122285] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.193914] env[62692]: DEBUG oslo_concurrency.lockutils [req-35c39032-75c5-4879-ac62-ea84968a9957 req-e1848722-46d5-4c73-b588-aedd57fcfaa2 service nova] Releasing lock "refresh_cache-628cc505-3edf-4066-91be-da009ebcf219" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.193914] env[62692]: DEBUG oslo_concurrency.lockutils [req-0fed755b-3004-41dc-8695-a8b43b602f0c req-6df88dc2-1371-4311-a703-54beb3a717bb service nova] Acquired lock "refresh_cache-628cc505-3edf-4066-91be-da009ebcf219" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.193914] env[62692]: DEBUG nova.network.neutron [req-0fed755b-3004-41dc-8695-a8b43b602f0c req-6df88dc2-1371-4311-a703-54beb3a717bb service nova] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Refreshing network info cache for port 30e027a7-86b8-4669-91fe-1dbd40ae54a5 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 786.346084] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b956b727-674f-4fdf-8fef-9624abd36e0c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.353657] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee2ec717-296c-417d-9bec-6934da589e96 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.385496] env[62692]: DEBUG nova.network.neutron [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Successfully created port: 6f540034-8dda-4def-adb2-0cdba3268510 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 786.387960] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05cd4f1d-386d-4380-8ef6-db1a1e4aa343 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.395211] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-253ca3e5-28c2-4571-a877-8524a92e015a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.409211] env[62692]: DEBUG nova.compute.manager [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 786.411842] env[62692]: DEBUG nova.compute.provider_tree [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 786.526031] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.526031] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 786.526031] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.526031] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.526290] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 786.526290] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cff81940-4e4b-4a16-8f72-5aa92b565e26 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.529751] env[62692]: DEBUG oslo_vmware.api [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 786.529751] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52c20c30-e832-b59b-bd66-1df4b5c3a46f" [ 786.529751] env[62692]: _type = "Task" [ 786.529751] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.538834] env[62692]: DEBUG oslo_vmware.api [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52c20c30-e832-b59b-bd66-1df4b5c3a46f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.560488] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140916, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.645534} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.562864] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Copied Virtual Disk [datastore1] vmware_temp/9657360f-fe08-490e-bf57-166bd1412e9f/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/tmp-sparse.vmdk to [datastore1] vmware_temp/9657360f-fe08-490e-bf57-166bd1412e9f/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 786.562864] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Deleting the datastore file [datastore1] vmware_temp/9657360f-fe08-490e-bf57-166bd1412e9f/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/tmp-sparse.vmdk {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 786.562864] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7f1a2531-2690-452e-89ed-8d266fd3659e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.568467] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 786.568467] env[62692]: value = "task-1140926" [ 786.568467] env[62692]: _type = "Task" [ 786.568467] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.577160] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140926, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.939043] env[62692]: ERROR nova.scheduler.client.report [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [req-c881128b-f23b-4ca4-b287-b110ca3e2fcf] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 470758a8-5efa-4875-aad5-f512a727752f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c881128b-f23b-4ca4-b287-b110ca3e2fcf"}]}: nova.exception.PortBindingFailed: Binding failed for port 8f492038-90fc-40ae-9f44-214df7488a50, please check neutron logs for more information. [ 786.959199] env[62692]: DEBUG nova.scheduler.client.report [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Refreshing inventories for resource provider 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 786.977030] env[62692]: DEBUG nova.scheduler.client.report [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Updating ProviderTree inventory for provider 470758a8-5efa-4875-aad5-f512a727752f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 786.977030] env[62692]: DEBUG nova.compute.provider_tree [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 786.988883] env[62692]: DEBUG nova.scheduler.client.report [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Refreshing aggregate associations for resource provider 470758a8-5efa-4875-aad5-f512a727752f, aggregates: None {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 787.010459] env[62692]: DEBUG nova.scheduler.client.report [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Refreshing trait associations for resource provider 470758a8-5efa-4875-aad5-f512a727752f, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 787.018263] env[62692]: DEBUG nova.network.neutron [req-0fed755b-3004-41dc-8695-a8b43b602f0c req-6df88dc2-1371-4311-a703-54beb3a717bb service nova] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Updated VIF entry in instance network info cache for port 30e027a7-86b8-4669-91fe-1dbd40ae54a5. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 787.019333] env[62692]: DEBUG nova.network.neutron [req-0fed755b-3004-41dc-8695-a8b43b602f0c req-6df88dc2-1371-4311-a703-54beb3a717bb service nova] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Updating instance_info_cache with network_info: [{"id": "30e027a7-86b8-4669-91fe-1dbd40ae54a5", "address": "fa:16:3e:2e:69:5e", "network": {"id": "7e7a1dcd-36ba-4a25-affa-6bd7fe895ade", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-508011613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9667b0b7abaa4bf0a7e54e9d540df728", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30e027a7-86", "ovs_interfaceid": "30e027a7-86b8-4669-91fe-1dbd40ae54a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.043303] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.043563] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 787.043770] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.080515] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140926, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.029443} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.083111] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 787.083508] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Moving file from [datastore1] vmware_temp/9657360f-fe08-490e-bf57-166bd1412e9f/81eb64fb-5a0c-47b7-9948-7dcb01b911bd to [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd. {{(pid=62692) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 787.083986] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-c93b128d-71c9-4a70-98f4-c69c46e73cbe {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.091942] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 787.091942] env[62692]: value = "task-1140927" [ 787.091942] env[62692]: _type = "Task" [ 787.091942] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.104316] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140927, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.339249] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fecd5c3c-b730-4cfe-8cbd-d360435bcac8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.346859] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aafb5fd9-c858-4e48-b839-1aa67195191a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.377819] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00b2034c-db18-4e2e-bc23-f82ae6d23de0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.385081] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71ec4701-c764-4e3b-bd1d-6b0a79bb9479 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.398473] env[62692]: DEBUG nova.compute.provider_tree [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 787.424104] env[62692]: DEBUG nova.compute.manager [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 787.451488] env[62692]: DEBUG nova.virt.hardware [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 787.451998] env[62692]: DEBUG nova.virt.hardware [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 787.451998] env[62692]: DEBUG nova.virt.hardware [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 787.452191] env[62692]: DEBUG nova.virt.hardware [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 787.452391] env[62692]: DEBUG nova.virt.hardware [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 787.452481] env[62692]: DEBUG nova.virt.hardware [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 787.452709] env[62692]: DEBUG nova.virt.hardware [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 787.452837] env[62692]: DEBUG nova.virt.hardware [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 787.453015] env[62692]: DEBUG nova.virt.hardware [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 787.453216] env[62692]: DEBUG nova.virt.hardware [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 787.453393] env[62692]: DEBUG nova.virt.hardware [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 787.454258] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f6e4635-d22c-43f6-85c1-8b9f5768e64c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.462994] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69a8115f-78c9-40d0-9530-fd6087563af5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.522460] env[62692]: DEBUG oslo_concurrency.lockutils [req-0fed755b-3004-41dc-8695-a8b43b602f0c req-6df88dc2-1371-4311-a703-54beb3a717bb service nova] Releasing lock "refresh_cache-628cc505-3edf-4066-91be-da009ebcf219" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.601604] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140927, 'name': MoveDatastoreFile_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.934679] env[62692]: DEBUG nova.scheduler.client.report [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Updated inventory for provider 470758a8-5efa-4875-aad5-f512a727752f with generation 63 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 787.934965] env[62692]: DEBUG nova.compute.provider_tree [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Updating resource provider 470758a8-5efa-4875-aad5-f512a727752f generation from 63 to 64 during operation: update_inventory {{(pid=62692) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 787.935185] env[62692]: DEBUG nova.compute.provider_tree [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 787.990451] env[62692]: DEBUG nova.compute.manager [req-3029e9ed-984c-45d0-b450-3eb4d42fc321 req-c40af4d1-630d-481b-866c-da7290245d8a service nova] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Received event network-vif-plugged-6f540034-8dda-4def-adb2-0cdba3268510 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 787.990677] env[62692]: DEBUG oslo_concurrency.lockutils [req-3029e9ed-984c-45d0-b450-3eb4d42fc321 req-c40af4d1-630d-481b-866c-da7290245d8a service nova] Acquiring lock "5a08fa8a-f9fe-4879-bb7b-baa04097df6c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.991172] env[62692]: DEBUG oslo_concurrency.lockutils [req-3029e9ed-984c-45d0-b450-3eb4d42fc321 req-c40af4d1-630d-481b-866c-da7290245d8a service nova] Lock "5a08fa8a-f9fe-4879-bb7b-baa04097df6c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.991307] env[62692]: DEBUG oslo_concurrency.lockutils [req-3029e9ed-984c-45d0-b450-3eb4d42fc321 req-c40af4d1-630d-481b-866c-da7290245d8a service nova] Lock "5a08fa8a-f9fe-4879-bb7b-baa04097df6c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.991458] env[62692]: DEBUG nova.compute.manager [req-3029e9ed-984c-45d0-b450-3eb4d42fc321 req-c40af4d1-630d-481b-866c-da7290245d8a service nova] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] No waiting events found dispatching network-vif-plugged-6f540034-8dda-4def-adb2-0cdba3268510 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 787.991637] env[62692]: WARNING nova.compute.manager [req-3029e9ed-984c-45d0-b450-3eb4d42fc321 req-c40af4d1-630d-481b-866c-da7290245d8a service nova] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Received unexpected event network-vif-plugged-6f540034-8dda-4def-adb2-0cdba3268510 for instance with vm_state building and task_state spawning. [ 788.085309] env[62692]: DEBUG nova.network.neutron [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Successfully updated port: 6f540034-8dda-4def-adb2-0cdba3268510 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 788.103421] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140927, 'name': MoveDatastoreFile_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.400441] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Acquiring lock "6b2d3385-a9af-432b-a1e7-c91adb2e1f69" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.400713] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Lock "6b2d3385-a9af-432b-a1e7-c91adb2e1f69" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.441102] env[62692]: DEBUG oslo_concurrency.lockutils [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 3.042s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.442090] env[62692]: ERROR nova.compute.manager [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8f492038-90fc-40ae-9f44-214df7488a50, please check neutron logs for more information. [ 788.442090] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] Traceback (most recent call last): [ 788.442090] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 788.442090] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] self.driver.spawn(context, instance, image_meta, [ 788.442090] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 543, in spawn [ 788.442090] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] self._vmops.spawn(context, instance, image_meta, injected_files, [ 788.442090] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 788.442090] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] vm_ref = self.build_virtual_machine(instance, [ 788.442090] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 788.442090] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] vif_infos = vmwarevif.get_vif_info(self._session, [ 788.442090] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 788.442478] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] for vif in network_info: [ 788.442478] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 788.442478] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] return self._sync_wrapper(fn, *args, **kwargs) [ 788.442478] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 788.442478] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] self.wait() [ 788.442478] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 788.442478] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] self[:] = self._gt.wait() [ 788.442478] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 788.442478] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] return self._exit_event.wait() [ 788.442478] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 788.442478] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] result = hub.switch() [ 788.442478] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 788.442478] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] return self.greenlet.switch() [ 788.442799] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 788.442799] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] result = function(*args, **kwargs) [ 788.442799] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 788.442799] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] return func(*args, **kwargs) [ 788.442799] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 788.442799] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] raise e [ 788.442799] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 788.442799] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] nwinfo = self.network_api.allocate_for_instance( [ 788.442799] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 788.442799] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] created_port_ids = self._update_ports_for_instance( [ 788.442799] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 788.442799] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] with excutils.save_and_reraise_exception(): [ 788.442799] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 788.443459] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] self.force_reraise() [ 788.443459] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 788.443459] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] raise self.value [ 788.443459] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 788.443459] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] updated_port = self._update_port( [ 788.443459] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 788.443459] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] _ensure_no_port_binding_failure(port) [ 788.443459] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 788.443459] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] raise exception.PortBindingFailed(port_id=port['id']) [ 788.443459] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] nova.exception.PortBindingFailed: Binding failed for port 8f492038-90fc-40ae-9f44-214df7488a50, please check neutron logs for more information. [ 788.443459] env[62692]: ERROR nova.compute.manager [instance: 49ee893a-313e-42ed-b932-352d450e8645] [ 788.444011] env[62692]: DEBUG nova.compute.utils [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Binding failed for port 8f492038-90fc-40ae-9f44-214df7488a50, please check neutron logs for more information. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 788.444472] env[62692]: DEBUG nova.compute.manager [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Build of instance 49ee893a-313e-42ed-b932-352d450e8645 was re-scheduled: Binding failed for port 8f492038-90fc-40ae-9f44-214df7488a50, please check neutron logs for more information. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 788.444926] env[62692]: DEBUG nova.compute.manager [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Unplugging VIFs for instance {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 788.445174] env[62692]: DEBUG oslo_concurrency.lockutils [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Acquiring lock "refresh_cache-49ee893a-313e-42ed-b932-352d450e8645" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.445325] env[62692]: DEBUG oslo_concurrency.lockutils [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Acquired lock "refresh_cache-49ee893a-313e-42ed-b932-352d450e8645" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.445485] env[62692]: DEBUG nova.network.neutron [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 788.447326] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.522s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.449337] env[62692]: INFO nova.compute.claims [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 788.589239] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "refresh_cache-5a08fa8a-f9fe-4879-bb7b-baa04097df6c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.589405] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquired lock "refresh_cache-5a08fa8a-f9fe-4879-bb7b-baa04097df6c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.589568] env[62692]: DEBUG nova.network.neutron [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 788.605025] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140927, 'name': MoveDatastoreFile_Task, 'duration_secs': 1.058406} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.605146] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] File moved {{(pid=62692) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 788.605857] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Cleaning up location [datastore1] vmware_temp/9657360f-fe08-490e-bf57-166bd1412e9f {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 788.605857] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Deleting the datastore file [datastore1] vmware_temp/9657360f-fe08-490e-bf57-166bd1412e9f {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 788.605857] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-52767af0-53f6-44e7-992f-a4799ecc441a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.613551] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 788.613551] env[62692]: value = "task-1140928" [ 788.613551] env[62692]: _type = "Task" [ 788.613551] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.622249] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140928, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.036336] env[62692]: DEBUG nova.network.neutron [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 789.129794] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140928, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.045286} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.130072] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 789.130806] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f12b1f00-0691-4cf6-a00a-407fb89f5780 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.135779] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 789.135779] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52293280-36f1-3e26-70ff-b9de310cec55" [ 789.135779] env[62692]: _type = "Task" [ 789.135779] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.143351] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52293280-36f1-3e26-70ff-b9de310cec55, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.158085] env[62692]: DEBUG nova.network.neutron [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 789.203469] env[62692]: DEBUG nova.network.neutron [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.383915] env[62692]: DEBUG nova.network.neutron [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Updating instance_info_cache with network_info: [{"id": "6f540034-8dda-4def-adb2-0cdba3268510", "address": "fa:16:3e:6c:33:e3", "network": {"id": "0e06e04a-5213-4d06-b3f1-b2b8f437e6d0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1533469056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "266c39fe585d41079d0278ecd1d1c6a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "678ebbe4-4c53-4eaf-a689-93981310f37d", "external-id": "nsx-vlan-transportzone-443", "segmentation_id": 443, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f540034-8d", "ovs_interfaceid": "6f540034-8dda-4def-adb2-0cdba3268510", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.647780] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52293280-36f1-3e26-70ff-b9de310cec55, 'name': SearchDatastore_Task, 'duration_secs': 0.130761} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.650530] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.650823] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] abee88ff-5cf5-4bf1-91e0-93b19cf30046/abee88ff-5cf5-4bf1-91e0-93b19cf30046.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 789.651324] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.651521] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 789.651788] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b8582a13-a906-4c0b-9f0d-b6986902b10e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.653663] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-79957e64-6a95-46ab-acb0-7fff4bed8b94 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.659998] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 789.659998] env[62692]: value = "task-1140930" [ 789.659998] env[62692]: _type = "Task" [ 789.659998] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.670182] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140930, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.671636] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 789.671818] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 789.675106] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8fa7f77b-9464-4a87-aacf-1180a313c159 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.680391] env[62692]: DEBUG oslo_vmware.api [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Waiting for the task: (returnval){ [ 789.680391] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]525952c8-2af4-e527-d688-c3a1de7a3c03" [ 789.680391] env[62692]: _type = "Task" [ 789.680391] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.689532] env[62692]: DEBUG oslo_vmware.api [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]525952c8-2af4-e527-d688-c3a1de7a3c03, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.707210] env[62692]: DEBUG oslo_concurrency.lockutils [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Releasing lock "refresh_cache-49ee893a-313e-42ed-b932-352d450e8645" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.707210] env[62692]: DEBUG nova.compute.manager [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62692) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 789.707210] env[62692]: DEBUG nova.compute.manager [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 789.708297] env[62692]: DEBUG nova.network.neutron [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 789.725772] env[62692]: DEBUG nova.network.neutron [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 789.819279] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe19432d-bc75-4a2d-a38b-b369b2e55816 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.827337] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-705cb8a2-29cf-4d99-bde6-7cf93652036a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.868979] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c77184fa-19d7-4291-90f3-34da4dd6985e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.876889] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-962f6fb1-bb17-4108-8a44-69c7e2d9107c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.890715] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Releasing lock "refresh_cache-5a08fa8a-f9fe-4879-bb7b-baa04097df6c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.891070] env[62692]: DEBUG nova.compute.manager [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Instance network_info: |[{"id": "6f540034-8dda-4def-adb2-0cdba3268510", "address": "fa:16:3e:6c:33:e3", "network": {"id": "0e06e04a-5213-4d06-b3f1-b2b8f437e6d0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1533469056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "266c39fe585d41079d0278ecd1d1c6a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "678ebbe4-4c53-4eaf-a689-93981310f37d", "external-id": "nsx-vlan-transportzone-443", "segmentation_id": 443, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f540034-8d", "ovs_interfaceid": "6f540034-8dda-4def-adb2-0cdba3268510", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 789.891562] env[62692]: DEBUG nova.compute.provider_tree [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 789.893047] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6c:33:e3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '678ebbe4-4c53-4eaf-a689-93981310f37d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6f540034-8dda-4def-adb2-0cdba3268510', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 789.900434] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Creating folder: Project (266c39fe585d41079d0278ecd1d1c6a4). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 789.901971] env[62692]: DEBUG nova.scheduler.client.report [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 789.904831] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-91933861-ef97-4103-b112-a2f0aed996d9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.915556] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Created folder: Project (266c39fe585d41079d0278ecd1d1c6a4) in parent group-v248868. [ 789.915760] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Creating folder: Instances. Parent ref: group-v248894. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 789.916633] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a2909a7b-009a-49f6-99b5-84ce0f7b33ff {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.927097] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Created folder: Instances in parent group-v248894. [ 789.927404] env[62692]: DEBUG oslo.service.loopingcall [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 789.927611] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 789.927838] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a7355815-bd84-470f-8e93-b80bb336f7f8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.947536] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 789.947536] env[62692]: value = "task-1140933" [ 789.947536] env[62692]: _type = "Task" [ 789.947536] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.956592] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140933, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.026525] env[62692]: DEBUG nova.compute.manager [req-edb1b977-5101-4905-9b36-11aab3bc3ad5 req-7029a0a3-7b62-4cfa-be12-11fbe3d0fb77 service nova] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Received event network-changed-6f540034-8dda-4def-adb2-0cdba3268510 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 790.026772] env[62692]: DEBUG nova.compute.manager [req-edb1b977-5101-4905-9b36-11aab3bc3ad5 req-7029a0a3-7b62-4cfa-be12-11fbe3d0fb77 service nova] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Refreshing instance network info cache due to event network-changed-6f540034-8dda-4def-adb2-0cdba3268510. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 790.026968] env[62692]: DEBUG oslo_concurrency.lockutils [req-edb1b977-5101-4905-9b36-11aab3bc3ad5 req-7029a0a3-7b62-4cfa-be12-11fbe3d0fb77 service nova] Acquiring lock "refresh_cache-5a08fa8a-f9fe-4879-bb7b-baa04097df6c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.027152] env[62692]: DEBUG oslo_concurrency.lockutils [req-edb1b977-5101-4905-9b36-11aab3bc3ad5 req-7029a0a3-7b62-4cfa-be12-11fbe3d0fb77 service nova] Acquired lock "refresh_cache-5a08fa8a-f9fe-4879-bb7b-baa04097df6c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.027359] env[62692]: DEBUG nova.network.neutron [req-edb1b977-5101-4905-9b36-11aab3bc3ad5 req-7029a0a3-7b62-4cfa-be12-11fbe3d0fb77 service nova] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Refreshing network info cache for port 6f540034-8dda-4def-adb2-0cdba3268510 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 790.173116] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140930, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.190987] env[62692]: DEBUG oslo_vmware.api [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]525952c8-2af4-e527-d688-c3a1de7a3c03, 'name': SearchDatastore_Task, 'duration_secs': 0.014224} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.191894] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5314c115-5bec-49ed-b02b-da34a8f5b902 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.198145] env[62692]: DEBUG oslo_vmware.api [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Waiting for the task: (returnval){ [ 790.198145] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52b586db-ab4a-7c6a-64e4-f607044d765f" [ 790.198145] env[62692]: _type = "Task" [ 790.198145] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.209239] env[62692]: DEBUG oslo_vmware.api [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52b586db-ab4a-7c6a-64e4-f607044d765f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.229485] env[62692]: DEBUG nova.network.neutron [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.407955] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.961s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.409338] env[62692]: DEBUG nova.compute.manager [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 790.411714] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.188s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.412854] env[62692]: INFO nova.compute.claims [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 790.461618] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140933, 'name': CreateVM_Task, 'duration_secs': 0.457036} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.461770] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 790.462447] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.462612] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.463012] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 790.463187] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3445806-7d38-49f8-bc7a-6250618133b4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.467903] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 790.467903] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5284263f-33d1-3eb8-f6ef-38db0b51903f" [ 790.467903] env[62692]: _type = "Task" [ 790.467903] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.475725] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5284263f-33d1-3eb8-f6ef-38db0b51903f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.672500] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140930, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.765753} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.672759] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] abee88ff-5cf5-4bf1-91e0-93b19cf30046/abee88ff-5cf5-4bf1-91e0-93b19cf30046.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 790.672972] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 790.673236] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-440a0935-cc35-4001-8180-7e8cf173506d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.681933] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 790.681933] env[62692]: value = "task-1140934" [ 790.681933] env[62692]: _type = "Task" [ 790.681933] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.688821] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140934, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.707891] env[62692]: DEBUG oslo_vmware.api [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52b586db-ab4a-7c6a-64e4-f607044d765f, 'name': SearchDatastore_Task, 'duration_secs': 0.049228} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.708177] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.708438] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] c1818ed2-9e27-4738-bc17-98832318cf61/c1818ed2-9e27-4738-bc17-98832318cf61.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 790.708771] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.708959] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 790.709173] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3963eb5b-a89d-4a41-ab9f-f4d182843e27 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.711090] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6bf8f9f1-ffcc-41ab-a819-4657a63bf381 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.720020] env[62692]: DEBUG oslo_vmware.api [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Waiting for the task: (returnval){ [ 790.720020] env[62692]: value = "task-1140935" [ 790.720020] env[62692]: _type = "Task" [ 790.720020] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.720379] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 790.720545] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 790.725081] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2b81000-263b-4611-9ffb-9655afb367e0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.734456] env[62692]: INFO nova.compute.manager [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] [instance: 49ee893a-313e-42ed-b932-352d450e8645] Took 1.03 seconds to deallocate network for instance. [ 790.737482] env[62692]: DEBUG oslo_vmware.api [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Task: {'id': task-1140935, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.743228] env[62692]: DEBUG oslo_vmware.api [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 790.743228] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5264e938-6a11-8050-dbcd-36408343df3e" [ 790.743228] env[62692]: _type = "Task" [ 790.743228] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.754044] env[62692]: DEBUG oslo_vmware.api [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5264e938-6a11-8050-dbcd-36408343df3e, 'name': SearchDatastore_Task, 'duration_secs': 0.015089} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.757125] env[62692]: DEBUG nova.network.neutron [req-edb1b977-5101-4905-9b36-11aab3bc3ad5 req-7029a0a3-7b62-4cfa-be12-11fbe3d0fb77 service nova] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Updated VIF entry in instance network info cache for port 6f540034-8dda-4def-adb2-0cdba3268510. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 790.757491] env[62692]: DEBUG nova.network.neutron [req-edb1b977-5101-4905-9b36-11aab3bc3ad5 req-7029a0a3-7b62-4cfa-be12-11fbe3d0fb77 service nova] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Updating instance_info_cache with network_info: [{"id": "6f540034-8dda-4def-adb2-0cdba3268510", "address": "fa:16:3e:6c:33:e3", "network": {"id": "0e06e04a-5213-4d06-b3f1-b2b8f437e6d0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1533469056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "266c39fe585d41079d0278ecd1d1c6a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "678ebbe4-4c53-4eaf-a689-93981310f37d", "external-id": "nsx-vlan-transportzone-443", "segmentation_id": 443, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f540034-8d", "ovs_interfaceid": "6f540034-8dda-4def-adb2-0cdba3268510", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.759524] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fac3e201-07bd-4837-8340-01d9af9a57c9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.766215] env[62692]: DEBUG oslo_vmware.api [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 790.766215] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]527dd5d5-5cfc-d818-a6cf-f7c2aab711e0" [ 790.766215] env[62692]: _type = "Task" [ 790.766215] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.776651] env[62692]: DEBUG oslo_vmware.api [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]527dd5d5-5cfc-d818-a6cf-f7c2aab711e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.917722] env[62692]: DEBUG nova.compute.utils [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 790.921757] env[62692]: DEBUG nova.compute.manager [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 790.921757] env[62692]: DEBUG nova.network.neutron [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 790.966076] env[62692]: DEBUG nova.policy [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9e9f256c1ebf4b1ba966edb7b1df5f4f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '266c39fe585d41079d0278ecd1d1c6a4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 790.980327] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5284263f-33d1-3eb8-f6ef-38db0b51903f, 'name': SearchDatastore_Task, 'duration_secs': 0.008449} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.980613] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.980858] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 790.981076] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.195696] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140934, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.219457} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.195897] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 791.196765] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dca1d000-d691-46f1-947f-75fe55bd04dc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.224470] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Reconfiguring VM instance instance-00000023 to attach disk [datastore1] abee88ff-5cf5-4bf1-91e0-93b19cf30046/abee88ff-5cf5-4bf1-91e0-93b19cf30046.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 791.224928] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3ddc2271-abd9-4167-b5cd-48ff066a5946 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.258026] env[62692]: DEBUG oslo_vmware.api [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Task: {'id': task-1140935, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.259232] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 791.259232] env[62692]: value = "task-1140937" [ 791.259232] env[62692]: _type = "Task" [ 791.259232] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.260584] env[62692]: DEBUG oslo_concurrency.lockutils [req-edb1b977-5101-4905-9b36-11aab3bc3ad5 req-7029a0a3-7b62-4cfa-be12-11fbe3d0fb77 service nova] Releasing lock "refresh_cache-5a08fa8a-f9fe-4879-bb7b-baa04097df6c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.279899] env[62692]: DEBUG oslo_vmware.api [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]527dd5d5-5cfc-d818-a6cf-f7c2aab711e0, 'name': SearchDatastore_Task, 'duration_secs': 0.010791} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.280228] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.280519] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] fa7bb883-0065-4ecc-9eb5-8c5443344a2a/fa7bb883-0065-4ecc-9eb5-8c5443344a2a.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 791.282134] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.282134] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 791.282134] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-52aa2ca6-d9ad-4b96-9406-984433d960d7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.284145] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e5f440c0-85f1-498b-bc61-d4d2dbef7173 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.294636] env[62692]: DEBUG oslo_vmware.api [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 791.294636] env[62692]: value = "task-1140938" [ 791.294636] env[62692]: _type = "Task" [ 791.294636] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.300379] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 791.300573] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 791.301383] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7423c880-cee5-4e89-be9a-b9a318e583cf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.308331] env[62692]: DEBUG oslo_vmware.api [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140938, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.310125] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 791.310125] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52f07933-b9d4-4973-c35a-0604cfdcb9f7" [ 791.310125] env[62692]: _type = "Task" [ 791.310125] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.319715] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52f07933-b9d4-4973-c35a-0604cfdcb9f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.337457] env[62692]: DEBUG nova.network.neutron [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Successfully created port: cad44f99-9f63-4d6e-a084-4061f88e126f {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 791.423073] env[62692]: DEBUG nova.compute.manager [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 791.738193] env[62692]: DEBUG oslo_vmware.api [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Task: {'id': task-1140935, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.792343} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.741077] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] c1818ed2-9e27-4738-bc17-98832318cf61/c1818ed2-9e27-4738-bc17-98832318cf61.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 791.741379] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 791.742587] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-40dac10e-e43a-403e-be2f-11ccfbeabe97 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.754169] env[62692]: DEBUG oslo_vmware.api [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Waiting for the task: (returnval){ [ 791.754169] env[62692]: value = "task-1140939" [ 791.754169] env[62692]: _type = "Task" [ 791.754169] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.770490] env[62692]: DEBUG oslo_vmware.api [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Task: {'id': task-1140939, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.776620] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140937, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.777918] env[62692]: INFO nova.scheduler.client.report [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Deleted allocations for instance 49ee893a-313e-42ed-b932-352d450e8645 [ 791.809976] env[62692]: DEBUG oslo_vmware.api [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140938, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.823362] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52f07933-b9d4-4973-c35a-0604cfdcb9f7, 'name': SearchDatastore_Task, 'duration_secs': 0.058877} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.824786] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf8e465f-9733-498e-9052-19f11debc357 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.832562] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 791.832562] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5275942d-b1cd-9021-5710-159c278e3027" [ 791.832562] env[62692]: _type = "Task" [ 791.832562] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.843426] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5275942d-b1cd-9021-5710-159c278e3027, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.871784] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43b5a4bd-2949-4821-ba95-69a60451509d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.883936] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f30c26a-50d6-4511-9651-eaee8f53970c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.920305] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-159e998d-2863-4547-a1f6-72aad78b073a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.931013] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d75fb90-1a58-4dcb-ae75-c186e85b8abd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.954572] env[62692]: DEBUG nova.compute.provider_tree [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.266879] env[62692]: DEBUG oslo_vmware.api [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Task: {'id': task-1140939, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.294431} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.267555] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 792.268338] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e31447ec-23a6-4880-a59d-238b0b138a9d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.275496] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140937, 'name': ReconfigVM_Task, 'duration_secs': 0.881522} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.276066] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Reconfigured VM instance instance-00000023 to attach disk [datastore1] abee88ff-5cf5-4bf1-91e0-93b19cf30046/abee88ff-5cf5-4bf1-91e0-93b19cf30046.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 792.276683] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6bd86485-7fd8-4f78-a545-28a4cf5fe927 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.292747] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Reconfiguring VM instance instance-00000025 to attach disk [datastore1] c1818ed2-9e27-4738-bc17-98832318cf61/c1818ed2-9e27-4738-bc17-98832318cf61.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 792.293562] env[62692]: DEBUG oslo_concurrency.lockutils [None req-58a38f4f-f4ba-4431-94d3-8af0b04c9cef tempest-AttachInterfacesUnderV243Test-1139743903 tempest-AttachInterfacesUnderV243Test-1139743903-project-member] Lock "49ee893a-313e-42ed-b932-352d450e8645" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 144.672s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.293756] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae73261f-91d6-4bde-b161-1413b5638be5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.309838] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 792.309838] env[62692]: value = "task-1140940" [ 792.309838] env[62692]: _type = "Task" [ 792.309838] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.321906] env[62692]: DEBUG oslo_vmware.api [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Waiting for the task: (returnval){ [ 792.321906] env[62692]: value = "task-1140941" [ 792.321906] env[62692]: _type = "Task" [ 792.321906] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.330256] env[62692]: DEBUG oslo_vmware.api [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140938, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.727923} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.330496] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140940, 'name': Rename_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.335276] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] fa7bb883-0065-4ecc-9eb5-8c5443344a2a/fa7bb883-0065-4ecc-9eb5-8c5443344a2a.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 792.335405] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 792.336252] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d7fccdba-a8a7-450e-ad09-d7fcfff4d3bb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.349818] env[62692]: DEBUG oslo_vmware.api [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Task: {'id': task-1140941, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.355570] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5275942d-b1cd-9021-5710-159c278e3027, 'name': SearchDatastore_Task, 'duration_secs': 0.063626} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.355918] env[62692]: DEBUG oslo_vmware.api [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 792.355918] env[62692]: value = "task-1140942" [ 792.355918] env[62692]: _type = "Task" [ 792.355918] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.356237] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.356540] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 5a08fa8a-f9fe-4879-bb7b-baa04097df6c/5a08fa8a-f9fe-4879-bb7b-baa04097df6c.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 792.356881] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-50576e56-c623-4269-a51b-3efc21a5a47f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.368402] env[62692]: DEBUG oslo_vmware.api [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140942, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.369725] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 792.369725] env[62692]: value = "task-1140943" [ 792.369725] env[62692]: _type = "Task" [ 792.369725] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.377716] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140943, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.441826] env[62692]: DEBUG nova.compute.manager [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 792.459613] env[62692]: DEBUG nova.scheduler.client.report [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 792.473909] env[62692]: DEBUG nova.virt.hardware [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 792.474247] env[62692]: DEBUG nova.virt.hardware [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 792.474948] env[62692]: DEBUG nova.virt.hardware [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 792.474948] env[62692]: DEBUG nova.virt.hardware [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 792.474948] env[62692]: DEBUG nova.virt.hardware [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 792.475159] env[62692]: DEBUG nova.virt.hardware [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 792.475452] env[62692]: DEBUG nova.virt.hardware [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 792.475679] env[62692]: DEBUG nova.virt.hardware [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 792.476394] env[62692]: DEBUG nova.virt.hardware [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 792.476394] env[62692]: DEBUG nova.virt.hardware [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 792.476394] env[62692]: DEBUG nova.virt.hardware [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 792.477373] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56e11a4f-5d1c-46b9-9927-79bb59444181 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.487259] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83b600aa-1d19-44b2-a2a8-bba6eb0e0834 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.815021] env[62692]: DEBUG nova.compute.manager [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 792.825737] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140940, 'name': Rename_Task, 'duration_secs': 0.165422} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.826350] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 792.826764] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6e486cb2-b660-485f-8115-b54a80f100dd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.835348] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 792.835348] env[62692]: value = "task-1140944" [ 792.835348] env[62692]: _type = "Task" [ 792.835348] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.844160] env[62692]: DEBUG oslo_vmware.api [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Task: {'id': task-1140941, 'name': ReconfigVM_Task, 'duration_secs': 0.336212} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.845027] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Reconfigured VM instance instance-00000025 to attach disk [datastore1] c1818ed2-9e27-4738-bc17-98832318cf61/c1818ed2-9e27-4738-bc17-98832318cf61.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 792.845926] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c4465ef2-256d-4365-8ecc-6eb25b70fb80 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.852587] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140944, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.863440] env[62692]: DEBUG oslo_vmware.api [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Waiting for the task: (returnval){ [ 792.863440] env[62692]: value = "task-1140945" [ 792.863440] env[62692]: _type = "Task" [ 792.863440] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.870106] env[62692]: DEBUG oslo_vmware.api [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140942, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075151} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.871085] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 792.876918] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c936159-9b68-41af-9d1c-cb4f2d26b049 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.883742] env[62692]: DEBUG oslo_vmware.api [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Task: {'id': task-1140945, 'name': Rename_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.889875] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140943, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.913851] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Reconfiguring VM instance instance-00000024 to attach disk [datastore1] fa7bb883-0065-4ecc-9eb5-8c5443344a2a/fa7bb883-0065-4ecc-9eb5-8c5443344a2a.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 792.914581] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cc4e3d10-35d4-4380-bc92-4af5c3c40fdd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.937811] env[62692]: DEBUG oslo_vmware.api [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 792.937811] env[62692]: value = "task-1140946" [ 792.937811] env[62692]: _type = "Task" [ 792.937811] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.946659] env[62692]: DEBUG oslo_vmware.api [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140946, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.967764] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.556s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.968402] env[62692]: DEBUG nova.compute.manager [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 792.971344] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.904s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.976615] env[62692]: INFO nova.compute.claims [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 793.000950] env[62692]: DEBUG nova.compute.manager [req-258399f6-f1a8-477a-aed8-fa8f17f3c468 req-aa16c023-31a1-4b69-b21a-7ed9db698cc7 service nova] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Received event network-vif-plugged-cad44f99-9f63-4d6e-a084-4061f88e126f {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 793.001335] env[62692]: DEBUG oslo_concurrency.lockutils [req-258399f6-f1a8-477a-aed8-fa8f17f3c468 req-aa16c023-31a1-4b69-b21a-7ed9db698cc7 service nova] Acquiring lock "9b8e830c-61b7-4dd2-8324-d3a96eec1465-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.001653] env[62692]: DEBUG oslo_concurrency.lockutils [req-258399f6-f1a8-477a-aed8-fa8f17f3c468 req-aa16c023-31a1-4b69-b21a-7ed9db698cc7 service nova] Lock "9b8e830c-61b7-4dd2-8324-d3a96eec1465-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.001899] env[62692]: DEBUG oslo_concurrency.lockutils [req-258399f6-f1a8-477a-aed8-fa8f17f3c468 req-aa16c023-31a1-4b69-b21a-7ed9db698cc7 service nova] Lock "9b8e830c-61b7-4dd2-8324-d3a96eec1465-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.002712] env[62692]: DEBUG nova.compute.manager [req-258399f6-f1a8-477a-aed8-fa8f17f3c468 req-aa16c023-31a1-4b69-b21a-7ed9db698cc7 service nova] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] No waiting events found dispatching network-vif-plugged-cad44f99-9f63-4d6e-a084-4061f88e126f {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 793.002712] env[62692]: WARNING nova.compute.manager [req-258399f6-f1a8-477a-aed8-fa8f17f3c468 req-aa16c023-31a1-4b69-b21a-7ed9db698cc7 service nova] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Received unexpected event network-vif-plugged-cad44f99-9f63-4d6e-a084-4061f88e126f for instance with vm_state building and task_state spawning. [ 793.091212] env[62692]: DEBUG nova.network.neutron [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Successfully updated port: cad44f99-9f63-4d6e-a084-4061f88e126f {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 793.347647] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140944, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.350020] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.379143] env[62692]: DEBUG oslo_vmware.api [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Task: {'id': task-1140945, 'name': Rename_Task, 'duration_secs': 0.194705} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.379470] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 793.380328] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7d58ab54-6f0a-448c-a96d-845d39ce5467 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.385874] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140943, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.537891} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.386443] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 5a08fa8a-f9fe-4879-bb7b-baa04097df6c/5a08fa8a-f9fe-4879-bb7b-baa04097df6c.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 793.386443] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 793.386805] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ae50db84-ba5b-4af5-81a3-55d69b13fc63 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.390097] env[62692]: DEBUG oslo_vmware.api [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Waiting for the task: (returnval){ [ 793.390097] env[62692]: value = "task-1140948" [ 793.390097] env[62692]: _type = "Task" [ 793.390097] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.400471] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 793.400471] env[62692]: value = "task-1140949" [ 793.400471] env[62692]: _type = "Task" [ 793.400471] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.403752] env[62692]: DEBUG oslo_vmware.api [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Task: {'id': task-1140948, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.412464] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140949, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.452536] env[62692]: DEBUG oslo_vmware.api [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140946, 'name': ReconfigVM_Task, 'duration_secs': 0.32082} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.453023] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Reconfigured VM instance instance-00000024 to attach disk [datastore1] fa7bb883-0065-4ecc-9eb5-8c5443344a2a/fa7bb883-0065-4ecc-9eb5-8c5443344a2a.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 793.453862] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bd9f0444-b247-4bcc-b786-4f2b93b2c5e3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.463216] env[62692]: DEBUG oslo_vmware.api [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 793.463216] env[62692]: value = "task-1140950" [ 793.463216] env[62692]: _type = "Task" [ 793.463216] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.473051] env[62692]: DEBUG oslo_vmware.api [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140950, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.481709] env[62692]: DEBUG nova.compute.utils [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 793.485519] env[62692]: DEBUG nova.compute.manager [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 793.485745] env[62692]: DEBUG nova.network.neutron [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 793.556521] env[62692]: DEBUG nova.policy [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4e5adb2c66c46cd8d5c8dce2c1f1c5f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '98ff5abb993645cb9a3ab28585fa2190', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 793.596809] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "refresh_cache-9b8e830c-61b7-4dd2-8324-d3a96eec1465" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.596967] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquired lock "refresh_cache-9b8e830c-61b7-4dd2-8324-d3a96eec1465" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.597145] env[62692]: DEBUG nova.network.neutron [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 793.848573] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140944, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.901637] env[62692]: DEBUG oslo_vmware.api [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Task: {'id': task-1140948, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.913316] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140949, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0765} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.914173] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 793.914478] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31ec3024-cf10-4893-a802-aac1e76bef4d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.919367] env[62692]: DEBUG nova.network.neutron [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Successfully created port: 469cb44a-0d70-4493-a248-8c3336ebc7b6 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 793.939794] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Reconfiguring VM instance instance-00000026 to attach disk [datastore1] 5a08fa8a-f9fe-4879-bb7b-baa04097df6c/5a08fa8a-f9fe-4879-bb7b-baa04097df6c.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 793.939794] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e0196ad4-065d-4740-98e1-96e90a4827c8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.960830] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 793.960830] env[62692]: value = "task-1140951" [ 793.960830] env[62692]: _type = "Task" [ 793.960830] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.972315] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140951, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.975427] env[62692]: DEBUG oslo_vmware.api [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140950, 'name': Rename_Task, 'duration_secs': 0.163048} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.975687] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 793.975911] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-99c31c9b-849c-404b-a859-82eebafb82e4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.983793] env[62692]: DEBUG oslo_vmware.api [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 793.983793] env[62692]: value = "task-1140952" [ 793.983793] env[62692]: _type = "Task" [ 793.983793] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.991687] env[62692]: DEBUG nova.compute.manager [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 793.997729] env[62692]: DEBUG oslo_vmware.api [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140952, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.148131] env[62692]: DEBUG nova.network.neutron [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 794.359931] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140944, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.388830] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d89c4e7-c31b-4661-ad3f-228ec5d5935c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.399224] env[62692]: DEBUG nova.network.neutron [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Updating instance_info_cache with network_info: [{"id": "cad44f99-9f63-4d6e-a084-4061f88e126f", "address": "fa:16:3e:81:2b:6e", "network": {"id": "0e06e04a-5213-4d06-b3f1-b2b8f437e6d0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1533469056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "266c39fe585d41079d0278ecd1d1c6a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "678ebbe4-4c53-4eaf-a689-93981310f37d", "external-id": "nsx-vlan-transportzone-443", "segmentation_id": 443, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcad44f99-9f", "ovs_interfaceid": "cad44f99-9f63-4d6e-a084-4061f88e126f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.400472] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2704f0d9-a38e-4cea-92a2-808a2191f945 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.411301] env[62692]: DEBUG oslo_vmware.api [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Task: {'id': task-1140948, 'name': PowerOnVM_Task, 'duration_secs': 0.515903} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.415214] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 794.415214] env[62692]: INFO nova.compute.manager [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Took 9.55 seconds to spawn the instance on the hypervisor. [ 794.415214] env[62692]: DEBUG nova.compute.manager [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 794.416026] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed529084-0663-452f-afc3-e39a027a5caf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.456994] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46e28686-aa72-4c42-ae57-ba99e7e66447 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.471935] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7996e72e-2440-4eae-9ee3-df3d33c9bb4c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.480312] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140951, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.492090] env[62692]: DEBUG nova.compute.provider_tree [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.507243] env[62692]: DEBUG oslo_vmware.api [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140952, 'name': PowerOnVM_Task, 'duration_secs': 0.506679} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.507243] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 794.507243] env[62692]: INFO nova.compute.manager [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Took 12.19 seconds to spawn the instance on the hypervisor. [ 794.507243] env[62692]: DEBUG nova.compute.manager [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 794.508046] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9e6f934-ccf1-4d02-b419-004c8cd257df {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.850783] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140944, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.909534] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Releasing lock "refresh_cache-9b8e830c-61b7-4dd2-8324-d3a96eec1465" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.909866] env[62692]: DEBUG nova.compute.manager [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Instance network_info: |[{"id": "cad44f99-9f63-4d6e-a084-4061f88e126f", "address": "fa:16:3e:81:2b:6e", "network": {"id": "0e06e04a-5213-4d06-b3f1-b2b8f437e6d0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1533469056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "266c39fe585d41079d0278ecd1d1c6a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "678ebbe4-4c53-4eaf-a689-93981310f37d", "external-id": "nsx-vlan-transportzone-443", "segmentation_id": 443, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcad44f99-9f", "ovs_interfaceid": "cad44f99-9f63-4d6e-a084-4061f88e126f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 794.910532] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:2b:6e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '678ebbe4-4c53-4eaf-a689-93981310f37d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cad44f99-9f63-4d6e-a084-4061f88e126f', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 794.923617] env[62692]: DEBUG oslo.service.loopingcall [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 794.923617] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 794.925132] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0afe1f8c-9339-4de2-bd19-b7e20989a185 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.946601] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 794.946601] env[62692]: value = "task-1140953" [ 794.946601] env[62692]: _type = "Task" [ 794.946601] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.955357] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140953, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.974547] env[62692]: INFO nova.compute.manager [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Took 44.52 seconds to build instance. [ 794.981708] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140951, 'name': ReconfigVM_Task, 'duration_secs': 0.95686} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.982403] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Reconfigured VM instance instance-00000026 to attach disk [datastore1] 5a08fa8a-f9fe-4879-bb7b-baa04097df6c/5a08fa8a-f9fe-4879-bb7b-baa04097df6c.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 794.983209] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9f5d4f00-23c0-45ba-8b12-dae1605dc629 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.992965] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 794.992965] env[62692]: value = "task-1140954" [ 794.992965] env[62692]: _type = "Task" [ 794.992965] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.999018] env[62692]: DEBUG nova.scheduler.client.report [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 795.009093] env[62692]: DEBUG nova.compute.manager [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 795.011120] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140954, 'name': Rename_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.026936] env[62692]: INFO nova.compute.manager [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Took 46.57 seconds to build instance. [ 795.045552] env[62692]: DEBUG nova.virt.hardware [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 795.045863] env[62692]: DEBUG nova.virt.hardware [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 795.046514] env[62692]: DEBUG nova.virt.hardware [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 795.046882] env[62692]: DEBUG nova.virt.hardware [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 795.047074] env[62692]: DEBUG nova.virt.hardware [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 795.047239] env[62692]: DEBUG nova.virt.hardware [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 795.047699] env[62692]: DEBUG nova.virt.hardware [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 795.047699] env[62692]: DEBUG nova.virt.hardware [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 795.047860] env[62692]: DEBUG nova.virt.hardware [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 795.048010] env[62692]: DEBUG nova.virt.hardware [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 795.048199] env[62692]: DEBUG nova.virt.hardware [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 795.049176] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52edd986-2669-41e4-a89c-86fd498e90dd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.059862] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-504004d9-dd4b-4c42-af5f-aba8c6373235 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.077871] env[62692]: DEBUG nova.compute.manager [req-7a583877-db2d-4a63-80f4-20d8a48094ae req-0dcb090a-55b5-4d89-844f-d2f90e5615a0 service nova] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Received event network-changed-cad44f99-9f63-4d6e-a084-4061f88e126f {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 795.078146] env[62692]: DEBUG nova.compute.manager [req-7a583877-db2d-4a63-80f4-20d8a48094ae req-0dcb090a-55b5-4d89-844f-d2f90e5615a0 service nova] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Refreshing instance network info cache due to event network-changed-cad44f99-9f63-4d6e-a084-4061f88e126f. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 795.078315] env[62692]: DEBUG oslo_concurrency.lockutils [req-7a583877-db2d-4a63-80f4-20d8a48094ae req-0dcb090a-55b5-4d89-844f-d2f90e5615a0 service nova] Acquiring lock "refresh_cache-9b8e830c-61b7-4dd2-8324-d3a96eec1465" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.078581] env[62692]: DEBUG oslo_concurrency.lockutils [req-7a583877-db2d-4a63-80f4-20d8a48094ae req-0dcb090a-55b5-4d89-844f-d2f90e5615a0 service nova] Acquired lock "refresh_cache-9b8e830c-61b7-4dd2-8324-d3a96eec1465" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.078691] env[62692]: DEBUG nova.network.neutron [req-7a583877-db2d-4a63-80f4-20d8a48094ae req-0dcb090a-55b5-4d89-844f-d2f90e5615a0 service nova] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Refreshing network info cache for port cad44f99-9f63-4d6e-a084-4061f88e126f {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 795.355243] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140944, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.460406] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140953, 'name': CreateVM_Task, 'duration_secs': 0.439815} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.460406] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 795.460931] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.461101] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.461450] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 795.464280] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b33d5904-a795-4162-997e-94ed5e36af23 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.468480] env[62692]: DEBUG oslo_vmware.api [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 795.468480] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5225d7b2-82ba-8e5b-63a0-a016c3c29867" [ 795.468480] env[62692]: _type = "Task" [ 795.468480] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.479839] env[62692]: DEBUG oslo_vmware.api [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5225d7b2-82ba-8e5b-63a0-a016c3c29867, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.486671] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f79b454b-ad58-45f6-b78a-9c4c80588cc7 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Lock "c1818ed2-9e27-4738-bc17-98832318cf61" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.418s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.507248] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.536s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.507832] env[62692]: DEBUG nova.compute.manager [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 795.510793] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140954, 'name': Rename_Task, 'duration_secs': 0.198874} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.511063] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.157s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.512755] env[62692]: INFO nova.compute.claims [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 795.515815] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 795.516309] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3289c5d8-e9f9-45a7-a34b-f7a03ce6cda9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.525938] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 795.525938] env[62692]: value = "task-1140956" [ 795.525938] env[62692]: _type = "Task" [ 795.525938] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.529510] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab996d5d-5d0e-40c3-ad1a-50f0e267b310 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "fa7bb883-0065-4ecc-9eb5-8c5443344a2a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.929s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.542024] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140956, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.854575] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140944, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.944679] env[62692]: DEBUG nova.compute.manager [None req-63343f95-58f6-404e-83cb-0b07a3ec38d6 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 795.945601] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-228e17d0-b50d-4a03-aa13-b1628a085ac2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.979581] env[62692]: DEBUG oslo_vmware.api [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5225d7b2-82ba-8e5b-63a0-a016c3c29867, 'name': SearchDatastore_Task, 'duration_secs': 0.012074} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.979878] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.980120] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 795.980358] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.980503] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.980677] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 795.980942] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e53334d5-2642-4250-b200-8aac697cee44 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.987034] env[62692]: DEBUG nova.compute.manager [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 795.993019] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 795.993019] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 795.994247] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-031efb33-ab5e-46e0-b63d-4a00e22deae2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.006635] env[62692]: DEBUG oslo_vmware.api [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 796.006635] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52cba789-1834-3a1f-0b2e-cb9bf5c1fa4d" [ 796.006635] env[62692]: _type = "Task" [ 796.006635] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.018216] env[62692]: DEBUG oslo_vmware.api [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52cba789-1834-3a1f-0b2e-cb9bf5c1fa4d, 'name': SearchDatastore_Task, 'duration_secs': 0.013305} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.019226] env[62692]: DEBUG nova.compute.utils [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 796.021697] env[62692]: DEBUG nova.compute.manager [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 796.021841] env[62692]: DEBUG nova.network.neutron [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 796.026019] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eca3889d-f1e4-4719-b65f-ec73de7ecffd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.031440] env[62692]: DEBUG nova.compute.manager [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 796.039280] env[62692]: DEBUG oslo_vmware.api [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 796.039280] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]526e0e56-b640-a327-e574-5f1ba4d31e34" [ 796.039280] env[62692]: _type = "Task" [ 796.039280] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.039405] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140956, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.048677] env[62692]: DEBUG oslo_vmware.api [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]526e0e56-b640-a327-e574-5f1ba4d31e34, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.081096] env[62692]: DEBUG nova.network.neutron [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Successfully updated port: 469cb44a-0d70-4493-a248-8c3336ebc7b6 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 796.083755] env[62692]: DEBUG nova.network.neutron [req-7a583877-db2d-4a63-80f4-20d8a48094ae req-0dcb090a-55b5-4d89-844f-d2f90e5615a0 service nova] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Updated VIF entry in instance network info cache for port cad44f99-9f63-4d6e-a084-4061f88e126f. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 796.084803] env[62692]: DEBUG nova.network.neutron [req-7a583877-db2d-4a63-80f4-20d8a48094ae req-0dcb090a-55b5-4d89-844f-d2f90e5615a0 service nova] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Updating instance_info_cache with network_info: [{"id": "cad44f99-9f63-4d6e-a084-4061f88e126f", "address": "fa:16:3e:81:2b:6e", "network": {"id": "0e06e04a-5213-4d06-b3f1-b2b8f437e6d0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1533469056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "266c39fe585d41079d0278ecd1d1c6a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "678ebbe4-4c53-4eaf-a689-93981310f37d", "external-id": "nsx-vlan-transportzone-443", "segmentation_id": 443, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcad44f99-9f", "ovs_interfaceid": "cad44f99-9f63-4d6e-a084-4061f88e126f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.114076] env[62692]: DEBUG nova.policy [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd3c6698b721144c9adac1fd0c9c7a073', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7db5465cb5db465b92dfa570572b6e80', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 796.224332] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Acquiring lock "c1818ed2-9e27-4738-bc17-98832318cf61" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.224484] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Lock "c1818ed2-9e27-4738-bc17-98832318cf61" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.224624] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Acquiring lock "c1818ed2-9e27-4738-bc17-98832318cf61-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.224810] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Lock "c1818ed2-9e27-4738-bc17-98832318cf61-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.224976] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Lock "c1818ed2-9e27-4738-bc17-98832318cf61-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.227249] env[62692]: INFO nova.compute.manager [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Terminating instance [ 796.229568] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Acquiring lock "refresh_cache-c1818ed2-9e27-4738-bc17-98832318cf61" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.229885] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Acquired lock "refresh_cache-c1818ed2-9e27-4738-bc17-98832318cf61" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.229885] env[62692]: DEBUG nova.network.neutron [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 796.355787] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140944, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.462237] env[62692]: INFO nova.compute.manager [None req-63343f95-58f6-404e-83cb-0b07a3ec38d6 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] instance snapshotting [ 796.463031] env[62692]: DEBUG nova.objects.instance [None req-63343f95-58f6-404e-83cb-0b07a3ec38d6 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Lazy-loading 'flavor' on Instance uuid c1818ed2-9e27-4738-bc17-98832318cf61 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 796.530715] env[62692]: DEBUG nova.compute.manager [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 796.532721] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.557680] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140956, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.563957] env[62692]: DEBUG oslo_vmware.api [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]526e0e56-b640-a327-e574-5f1ba4d31e34, 'name': SearchDatastore_Task, 'duration_secs': 0.012941} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.564986] env[62692]: DEBUG nova.network.neutron [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Successfully created port: c77af962-553d-4504-825e-78a6689aee79 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 796.567208] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.567563] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 9b8e830c-61b7-4dd2-8324-d3a96eec1465/9b8e830c-61b7-4dd2-8324-d3a96eec1465.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 796.570372] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dbd69991-4b83-46c3-b96f-ce10e3fe597a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.574152] env[62692]: DEBUG oslo_concurrency.lockutils [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.584373] env[62692]: DEBUG oslo_vmware.api [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 796.584373] env[62692]: value = "task-1140957" [ 796.584373] env[62692]: _type = "Task" [ 796.584373] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.591199] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Acquiring lock "refresh_cache-105e6fd5-4eff-4a0c-9a4b-e6deade781d5" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.591199] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Acquired lock "refresh_cache-105e6fd5-4eff-4a0c-9a4b-e6deade781d5" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.591199] env[62692]: DEBUG nova.network.neutron [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 796.592616] env[62692]: DEBUG oslo_concurrency.lockutils [req-7a583877-db2d-4a63-80f4-20d8a48094ae req-0dcb090a-55b5-4d89-844f-d2f90e5615a0 service nova] Releasing lock "refresh_cache-9b8e830c-61b7-4dd2-8324-d3a96eec1465" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.605129] env[62692]: DEBUG oslo_vmware.api [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140957, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.763413] env[62692]: DEBUG nova.network.neutron [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 796.859721] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140944, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.970740] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cca73ae1-941d-41e0-8b68-87ce7e2302c0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.999415] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df503b30-8f67-4bfc-828d-aaf11bc34412 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.057912] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc0c863-d8f4-48b0-9460-47484f0e6271 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.067557] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140956, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.074447] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f97ce4c-f648-4cab-bcf2-1a28889e0dbd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.080733] env[62692]: INFO nova.compute.manager [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Rebuilding instance [ 797.136435] env[62692]: DEBUG nova.network.neutron [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.139046] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0701450b-8121-451f-b7f0-17d175e8915b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.151688] env[62692]: DEBUG oslo_vmware.api [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140957, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.161415] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f8c09b7-fddc-4d84-be57-590a7c41cc8b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.181145] env[62692]: DEBUG nova.compute.provider_tree [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 797.187404] env[62692]: DEBUG nova.compute.manager [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 797.188316] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-469caf0b-cbf2-46f8-8341-f9b7eb8afb51 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.195518] env[62692]: DEBUG nova.network.neutron [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 797.357727] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140944, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.463241] env[62692]: DEBUG nova.compute.manager [req-4f5699d0-f975-4716-8227-037b7ee953a9 req-afae61e1-4e83-473f-8080-b43726c093a2 service nova] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Received event network-vif-plugged-469cb44a-0d70-4493-a248-8c3336ebc7b6 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 797.463501] env[62692]: DEBUG oslo_concurrency.lockutils [req-4f5699d0-f975-4716-8227-037b7ee953a9 req-afae61e1-4e83-473f-8080-b43726c093a2 service nova] Acquiring lock "105e6fd5-4eff-4a0c-9a4b-e6deade781d5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.463773] env[62692]: DEBUG oslo_concurrency.lockutils [req-4f5699d0-f975-4716-8227-037b7ee953a9 req-afae61e1-4e83-473f-8080-b43726c093a2 service nova] Lock "105e6fd5-4eff-4a0c-9a4b-e6deade781d5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.463956] env[62692]: DEBUG oslo_concurrency.lockutils [req-4f5699d0-f975-4716-8227-037b7ee953a9 req-afae61e1-4e83-473f-8080-b43726c093a2 service nova] Lock "105e6fd5-4eff-4a0c-9a4b-e6deade781d5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.464349] env[62692]: DEBUG nova.compute.manager [req-4f5699d0-f975-4716-8227-037b7ee953a9 req-afae61e1-4e83-473f-8080-b43726c093a2 service nova] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] No waiting events found dispatching network-vif-plugged-469cb44a-0d70-4493-a248-8c3336ebc7b6 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 797.464410] env[62692]: WARNING nova.compute.manager [req-4f5699d0-f975-4716-8227-037b7ee953a9 req-afae61e1-4e83-473f-8080-b43726c093a2 service nova] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Received unexpected event network-vif-plugged-469cb44a-0d70-4493-a248-8c3336ebc7b6 for instance with vm_state building and task_state spawning. [ 797.464663] env[62692]: DEBUG nova.compute.manager [req-4f5699d0-f975-4716-8227-037b7ee953a9 req-afae61e1-4e83-473f-8080-b43726c093a2 service nova] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Received event network-changed-469cb44a-0d70-4493-a248-8c3336ebc7b6 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 797.464789] env[62692]: DEBUG nova.compute.manager [req-4f5699d0-f975-4716-8227-037b7ee953a9 req-afae61e1-4e83-473f-8080-b43726c093a2 service nova] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Refreshing instance network info cache due to event network-changed-469cb44a-0d70-4493-a248-8c3336ebc7b6. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 797.464962] env[62692]: DEBUG oslo_concurrency.lockutils [req-4f5699d0-f975-4716-8227-037b7ee953a9 req-afae61e1-4e83-473f-8080-b43726c093a2 service nova] Acquiring lock "refresh_cache-105e6fd5-4eff-4a0c-9a4b-e6deade781d5" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.517138] env[62692]: DEBUG nova.network.neutron [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Updating instance_info_cache with network_info: [{"id": "469cb44a-0d70-4493-a248-8c3336ebc7b6", "address": "fa:16:3e:53:00:df", "network": {"id": "a74c4d30-3568-4e50-98f5-d2fcc05ec728", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.15", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "989968e8796a4a70a4dbd045418a475f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap469cb44a-0d", "ovs_interfaceid": "469cb44a-0d70-4493-a248-8c3336ebc7b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.521016] env[62692]: DEBUG nova.compute.manager [None req-63343f95-58f6-404e-83cb-0b07a3ec38d6 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Instance disappeared during snapshot {{(pid=62692) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 797.550301] env[62692]: DEBUG nova.compute.manager [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 797.557557] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140956, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.577579] env[62692]: DEBUG nova.virt.hardware [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 797.578108] env[62692]: DEBUG nova.virt.hardware [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 797.578432] env[62692]: DEBUG nova.virt.hardware [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 797.578937] env[62692]: DEBUG nova.virt.hardware [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 797.579266] env[62692]: DEBUG nova.virt.hardware [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 797.579567] env[62692]: DEBUG nova.virt.hardware [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 797.579896] env[62692]: DEBUG nova.virt.hardware [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 797.580189] env[62692]: DEBUG nova.virt.hardware [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 797.580474] env[62692]: DEBUG nova.virt.hardware [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 797.580858] env[62692]: DEBUG nova.virt.hardware [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 797.582430] env[62692]: DEBUG nova.virt.hardware [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 797.582430] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc7a135a-57c1-4bd2-b3b2-c770f70a18f1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.595942] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3838dc0-787d-44cc-8391-069cab60bec0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.607043] env[62692]: DEBUG oslo_vmware.api [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140957, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.615} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.616578] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 9b8e830c-61b7-4dd2-8324-d3a96eec1465/9b8e830c-61b7-4dd2-8324-d3a96eec1465.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 797.616874] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 797.617327] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-37c650a1-a06c-46a9-a2f8-e742080d9f6a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.627859] env[62692]: DEBUG oslo_vmware.api [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 797.627859] env[62692]: value = "task-1140959" [ 797.627859] env[62692]: _type = "Task" [ 797.627859] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.638980] env[62692]: DEBUG oslo_vmware.api [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140959, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.643257] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Releasing lock "refresh_cache-c1818ed2-9e27-4738-bc17-98832318cf61" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.643716] env[62692]: DEBUG nova.compute.manager [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 797.644089] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 797.644835] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3503d0dc-12b2-46c6-982b-a55d84f0167a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.653494] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 797.654065] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e5b15159-f2e4-4bef-a7f1-ece9bb76b5e4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.665690] env[62692]: DEBUG oslo_vmware.api [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Waiting for the task: (returnval){ [ 797.665690] env[62692]: value = "task-1140960" [ 797.665690] env[62692]: _type = "Task" [ 797.665690] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.676578] env[62692]: DEBUG oslo_vmware.api [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Task: {'id': task-1140960, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.687247] env[62692]: DEBUG nova.scheduler.client.report [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 797.700025] env[62692]: DEBUG nova.compute.manager [None req-63343f95-58f6-404e-83cb-0b07a3ec38d6 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Found 0 images (rotation: 2) {{(pid=62692) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 797.706997] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 797.708028] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-792ff22a-39a7-42f0-8e6c-b9c9e615ea8e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.716996] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 797.716996] env[62692]: value = "task-1140961" [ 797.716996] env[62692]: _type = "Task" [ 797.716996] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.728729] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140961, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.861524] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140944, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.023954] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Releasing lock "refresh_cache-105e6fd5-4eff-4a0c-9a4b-e6deade781d5" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.023954] env[62692]: DEBUG nova.compute.manager [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Instance network_info: |[{"id": "469cb44a-0d70-4493-a248-8c3336ebc7b6", "address": "fa:16:3e:53:00:df", "network": {"id": "a74c4d30-3568-4e50-98f5-d2fcc05ec728", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.15", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "989968e8796a4a70a4dbd045418a475f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap469cb44a-0d", "ovs_interfaceid": "469cb44a-0d70-4493-a248-8c3336ebc7b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 798.024133] env[62692]: DEBUG oslo_concurrency.lockutils [req-4f5699d0-f975-4716-8227-037b7ee953a9 req-afae61e1-4e83-473f-8080-b43726c093a2 service nova] Acquired lock "refresh_cache-105e6fd5-4eff-4a0c-9a4b-e6deade781d5" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.024133] env[62692]: DEBUG nova.network.neutron [req-4f5699d0-f975-4716-8227-037b7ee953a9 req-afae61e1-4e83-473f-8080-b43726c093a2 service nova] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Refreshing network info cache for port 469cb44a-0d70-4493-a248-8c3336ebc7b6 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 798.025095] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:53:00:df', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd8383707-f093-40a7-a5ba-31b0e07cac45', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '469cb44a-0d70-4493-a248-8c3336ebc7b6', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 798.033219] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Creating folder: Project (98ff5abb993645cb9a3ab28585fa2190). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 798.036868] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-15f9e956-1755-4d96-9a57-3b5d17a13ecd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.050193] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140956, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.052717] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Created folder: Project (98ff5abb993645cb9a3ab28585fa2190) in parent group-v248868. [ 798.053099] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Creating folder: Instances. Parent ref: group-v248899. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 798.053443] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f3d7a7b8-2d28-4afe-a301-9fdf30d95b35 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.066275] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Created folder: Instances in parent group-v248899. [ 798.066275] env[62692]: DEBUG oslo.service.loopingcall [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 798.066275] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 798.066275] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1fc8f830-9a57-4771-9ad9-9f69ad52514c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.090231] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 798.090231] env[62692]: value = "task-1140964" [ 798.090231] env[62692]: _type = "Task" [ 798.090231] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.100895] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140964, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.139088] env[62692]: DEBUG oslo_vmware.api [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140959, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.128404} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.142933] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 798.144363] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13da8e04-478c-4d26-92d4-0c4c890b2e4e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.172905] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Reconfiguring VM instance instance-00000027 to attach disk [datastore2] 9b8e830c-61b7-4dd2-8324-d3a96eec1465/9b8e830c-61b7-4dd2-8324-d3a96eec1465.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 798.175726] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a6b91c4-82a9-41bd-8066-acc814acd96b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.195493] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.684s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.196261] env[62692]: DEBUG nova.compute.manager [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 798.199879] env[62692]: DEBUG oslo_concurrency.lockutils [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.216s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.201567] env[62692]: INFO nova.compute.claims [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 798.211733] env[62692]: DEBUG oslo_vmware.api [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Task: {'id': task-1140960, 'name': PowerOffVM_Task, 'duration_secs': 0.193225} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.214024] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 798.214024] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 798.214024] env[62692]: DEBUG oslo_vmware.api [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 798.214024] env[62692]: value = "task-1140965" [ 798.214024] env[62692]: _type = "Task" [ 798.214024] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.214024] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-509db75f-fc92-4ec3-b67c-9446bad7e2de {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.227929] env[62692]: DEBUG oslo_vmware.api [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140965, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.231785] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140961, 'name': PowerOffVM_Task, 'duration_secs': 0.22785} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.232156] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 798.232439] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 798.236023] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34ac5ff9-dd4e-487f-947d-fc62d95c2326 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.244988] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 798.245318] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-295c9d0b-2786-42ac-a84d-f74f7f67260e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.253149] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 798.253149] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 798.253149] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Deleting the datastore file [datastore1] c1818ed2-9e27-4738-bc17-98832318cf61 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 798.253149] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-245d5dfc-7a96-4f35-a53b-8feb64afd940 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.262410] env[62692]: DEBUG oslo_vmware.api [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Waiting for the task: (returnval){ [ 798.262410] env[62692]: value = "task-1140968" [ 798.262410] env[62692]: _type = "Task" [ 798.262410] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.272490] env[62692]: DEBUG oslo_vmware.api [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Task: {'id': task-1140968, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.359812] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140944, 'name': PowerOnVM_Task} progress is 80%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.427768] env[62692]: DEBUG nova.network.neutron [req-4f5699d0-f975-4716-8227-037b7ee953a9 req-afae61e1-4e83-473f-8080-b43726c093a2 service nova] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Updated VIF entry in instance network info cache for port 469cb44a-0d70-4493-a248-8c3336ebc7b6. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 798.427768] env[62692]: DEBUG nova.network.neutron [req-4f5699d0-f975-4716-8227-037b7ee953a9 req-afae61e1-4e83-473f-8080-b43726c093a2 service nova] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Updating instance_info_cache with network_info: [{"id": "469cb44a-0d70-4493-a248-8c3336ebc7b6", "address": "fa:16:3e:53:00:df", "network": {"id": "a74c4d30-3568-4e50-98f5-d2fcc05ec728", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.15", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "989968e8796a4a70a4dbd045418a475f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap469cb44a-0d", "ovs_interfaceid": "469cb44a-0d70-4493-a248-8c3336ebc7b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.552599] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140956, 'name': PowerOnVM_Task} progress is 91%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.600746] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140964, 'name': CreateVM_Task, 'duration_secs': 0.38367} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.600975] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 798.601797] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.602039] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.603996] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 798.603996] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a94df485-ea31-420a-8762-04a6eeafba5f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.609491] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Waiting for the task: (returnval){ [ 798.609491] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5280b5d4-6a83-6b00-05de-b0b9dd2d7279" [ 798.609491] env[62692]: _type = "Task" [ 798.609491] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.620327] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5280b5d4-6a83-6b00-05de-b0b9dd2d7279, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.694716] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 798.695580] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 798.695901] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Deleting the datastore file [datastore1] fa7bb883-0065-4ecc-9eb5-8c5443344a2a {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 798.696337] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-41a9125b-2be1-4666-8cbb-9d20c2457ce0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.709735] env[62692]: DEBUG nova.compute.utils [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 798.711107] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 798.711107] env[62692]: value = "task-1140969" [ 798.711107] env[62692]: _type = "Task" [ 798.711107] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.711378] env[62692]: DEBUG nova.compute.manager [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 798.711663] env[62692]: DEBUG nova.network.neutron [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 798.730151] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140969, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.734671] env[62692]: DEBUG oslo_vmware.api [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140965, 'name': ReconfigVM_Task, 'duration_secs': 0.390337} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.735221] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Reconfigured VM instance instance-00000027 to attach disk [datastore2] 9b8e830c-61b7-4dd2-8324-d3a96eec1465/9b8e830c-61b7-4dd2-8324-d3a96eec1465.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 798.735605] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c59ac210-75d9-40dd-9439-bb5993e52959 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.743888] env[62692]: DEBUG oslo_vmware.api [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 798.743888] env[62692]: value = "task-1140970" [ 798.743888] env[62692]: _type = "Task" [ 798.743888] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.753891] env[62692]: DEBUG oslo_vmware.api [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140970, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.758887] env[62692]: DEBUG nova.compute.manager [req-f604db0c-bc80-4741-b1b3-ed5c3a2682d1 req-64169856-1f36-428a-8890-09cada657feb service nova] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Received event network-vif-plugged-c77af962-553d-4504-825e-78a6689aee79 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 798.759119] env[62692]: DEBUG oslo_concurrency.lockutils [req-f604db0c-bc80-4741-b1b3-ed5c3a2682d1 req-64169856-1f36-428a-8890-09cada657feb service nova] Acquiring lock "bb547773-d176-4c8e-a0fa-a374d5050b1b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.759734] env[62692]: DEBUG oslo_concurrency.lockutils [req-f604db0c-bc80-4741-b1b3-ed5c3a2682d1 req-64169856-1f36-428a-8890-09cada657feb service nova] Lock "bb547773-d176-4c8e-a0fa-a374d5050b1b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.759893] env[62692]: DEBUG oslo_concurrency.lockutils [req-f604db0c-bc80-4741-b1b3-ed5c3a2682d1 req-64169856-1f36-428a-8890-09cada657feb service nova] Lock "bb547773-d176-4c8e-a0fa-a374d5050b1b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.760118] env[62692]: DEBUG nova.compute.manager [req-f604db0c-bc80-4741-b1b3-ed5c3a2682d1 req-64169856-1f36-428a-8890-09cada657feb service nova] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] No waiting events found dispatching network-vif-plugged-c77af962-553d-4504-825e-78a6689aee79 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 798.760303] env[62692]: WARNING nova.compute.manager [req-f604db0c-bc80-4741-b1b3-ed5c3a2682d1 req-64169856-1f36-428a-8890-09cada657feb service nova] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Received unexpected event network-vif-plugged-c77af962-553d-4504-825e-78a6689aee79 for instance with vm_state building and task_state spawning. [ 798.775122] env[62692]: DEBUG oslo_vmware.api [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Task: {'id': task-1140968, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.343235} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.775421] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 798.775622] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 798.776591] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 798.776591] env[62692]: INFO nova.compute.manager [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Took 1.13 seconds to destroy the instance on the hypervisor. [ 798.776591] env[62692]: DEBUG oslo.service.loopingcall [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 798.776795] env[62692]: DEBUG nova.compute.manager [-] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 798.776795] env[62692]: DEBUG nova.network.neutron [-] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 798.784667] env[62692]: DEBUG nova.policy [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd3c6698b721144c9adac1fd0c9c7a073', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7db5465cb5db465b92dfa570572b6e80', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 798.808209] env[62692]: DEBUG nova.network.neutron [-] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 798.861936] env[62692]: DEBUG oslo_vmware.api [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1140944, 'name': PowerOnVM_Task, 'duration_secs': 5.88551} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.862307] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 798.862608] env[62692]: INFO nova.compute.manager [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Took 19.05 seconds to spawn the instance on the hypervisor. [ 798.862715] env[62692]: DEBUG nova.compute.manager [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 798.863576] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23bd16ef-a987-40a1-b593-c7bdc92abc67 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.929674] env[62692]: DEBUG oslo_concurrency.lockutils [req-4f5699d0-f975-4716-8227-037b7ee953a9 req-afae61e1-4e83-473f-8080-b43726c093a2 service nova] Releasing lock "refresh_cache-105e6fd5-4eff-4a0c-9a4b-e6deade781d5" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.991365] env[62692]: DEBUG nova.network.neutron [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Successfully updated port: c77af962-553d-4504-825e-78a6689aee79 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 799.054819] env[62692]: DEBUG oslo_vmware.api [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140956, 'name': PowerOnVM_Task, 'duration_secs': 3.04919} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.059665] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 799.059665] env[62692]: INFO nova.compute.manager [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Took 11.64 seconds to spawn the instance on the hypervisor. [ 799.059665] env[62692]: DEBUG nova.compute.manager [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 799.060330] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68861303-c4e2-4626-8512-cf933c29d1c2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.128480] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5280b5d4-6a83-6b00-05de-b0b9dd2d7279, 'name': SearchDatastore_Task, 'duration_secs': 0.03025} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.128813] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.129315] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 799.129315] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.129464] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.129614] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 799.129845] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3d40f442-85f8-4592-961e-a08e6d25a664 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.141433] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 799.141622] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 799.142611] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-633ce300-085d-44ed-a01a-00d5de7b5489 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.151868] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Waiting for the task: (returnval){ [ 799.151868] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5280342a-de7b-6814-f8a5-8c9866348743" [ 799.151868] env[62692]: _type = "Task" [ 799.151868] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.160638] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5280342a-de7b-6814-f8a5-8c9866348743, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.161474] env[62692]: DEBUG nova.network.neutron [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Successfully created port: 920fbe7b-7ba1-4d33-bf4f-d96609ed6640 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 799.215274] env[62692]: DEBUG nova.compute.manager [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 799.228635] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140969, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.260353] env[62692]: DEBUG oslo_vmware.api [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140970, 'name': Rename_Task, 'duration_secs': 0.170529} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.260682] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 799.260966] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ba2b421a-0faf-4c2a-8b44-4164814769e4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.269566] env[62692]: DEBUG oslo_vmware.api [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 799.269566] env[62692]: value = "task-1140971" [ 799.269566] env[62692]: _type = "Task" [ 799.269566] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.280139] env[62692]: DEBUG oslo_vmware.api [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140971, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.312902] env[62692]: DEBUG nova.network.neutron [-] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.390083] env[62692]: INFO nova.compute.manager [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Took 52.93 seconds to build instance. [ 799.493991] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "refresh_cache-bb547773-d176-4c8e-a0fa-a374d5050b1b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.494658] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquired lock "refresh_cache-bb547773-d176-4c8e-a0fa-a374d5050b1b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.496099] env[62692]: DEBUG nova.network.neutron [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 799.590122] env[62692]: INFO nova.compute.manager [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Took 47.20 seconds to build instance. [ 799.664512] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5280342a-de7b-6814-f8a5-8c9866348743, 'name': SearchDatastore_Task, 'duration_secs': 0.040634} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.665504] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48c57623-9a91-4087-811f-9d1fa999446d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.678226] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Waiting for the task: (returnval){ [ 799.678226] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52c4aa53-5dc8-95ae-b85e-02e7dcb17d2d" [ 799.678226] env[62692]: _type = "Task" [ 799.678226] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.688873] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52c4aa53-5dc8-95ae-b85e-02e7dcb17d2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.704702] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9014dbf-719c-4c09-af6b-e6b6c0492914 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.713240] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f52f562a-c646-4237-bede-1f5ead550060 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.756863] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c661135-86f9-4d39-8553-1c21ed450d7a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.765755] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140969, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.68032} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.770260] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 799.770260] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 799.770260] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 799.771603] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a059bcd8-68d9-4aea-ba62-4e2db0421b72 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.784280] env[62692]: DEBUG oslo_vmware.api [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140971, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.793294] env[62692]: DEBUG nova.compute.provider_tree [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 799.816832] env[62692]: INFO nova.compute.manager [-] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Took 1.04 seconds to deallocate network for instance. [ 799.892371] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29063f55-62cc-4f54-addc-43169afecc8e tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lock "abee88ff-5cf5-4bf1-91e0-93b19cf30046" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 149.042s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.053597] env[62692]: DEBUG nova.network.neutron [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 800.091776] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c157510e-8114-4004-9db7-ba72ec343162 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "5a08fa8a-f9fe-4879-bb7b-baa04097df6c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 143.239s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.192108] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52c4aa53-5dc8-95ae-b85e-02e7dcb17d2d, 'name': SearchDatastore_Task, 'duration_secs': 0.012832} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.192515] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.192778] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 105e6fd5-4eff-4a0c-9a4b-e6deade781d5/105e6fd5-4eff-4a0c-9a4b-e6deade781d5.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 800.193056] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-02bf0cb3-d212-418a-ba4b-f3f6c2a5488f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.201323] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Waiting for the task: (returnval){ [ 800.201323] env[62692]: value = "task-1140973" [ 800.201323] env[62692]: _type = "Task" [ 800.201323] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.211830] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Task: {'id': task-1140973, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.258444] env[62692]: DEBUG nova.compute.manager [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 800.292858] env[62692]: DEBUG oslo_vmware.api [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1140971, 'name': PowerOnVM_Task, 'duration_secs': 0.593891} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.292858] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 800.292858] env[62692]: INFO nova.compute.manager [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Took 7.85 seconds to spawn the instance on the hypervisor. [ 800.292858] env[62692]: DEBUG nova.compute.manager [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 800.292858] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e4286b7-2689-4be5-8ae4-d732bc0d50b0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.298020] env[62692]: DEBUG nova.network.neutron [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Updating instance_info_cache with network_info: [{"id": "c77af962-553d-4504-825e-78a6689aee79", "address": "fa:16:3e:9a:01:cc", "network": {"id": "17e72c3b-603f-4305-96ac-89a9b4c9c7e9", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-123501993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7db5465cb5db465b92dfa570572b6e80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc77af962-55", "ovs_interfaceid": "c77af962-553d-4504-825e-78a6689aee79", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.302146] env[62692]: DEBUG nova.virt.hardware [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 800.302146] env[62692]: DEBUG nova.virt.hardware [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 800.302146] env[62692]: DEBUG nova.virt.hardware [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 800.302368] env[62692]: DEBUG nova.virt.hardware [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 800.302368] env[62692]: DEBUG nova.virt.hardware [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 800.302368] env[62692]: DEBUG nova.virt.hardware [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 800.302368] env[62692]: DEBUG nova.virt.hardware [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 800.302368] env[62692]: DEBUG nova.virt.hardware [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 800.302528] env[62692]: DEBUG nova.virt.hardware [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 800.302615] env[62692]: DEBUG nova.virt.hardware [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 800.302763] env[62692]: DEBUG nova.virt.hardware [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 800.306162] env[62692]: DEBUG nova.scheduler.client.report [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 800.309398] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85fb503a-b546-4ef9-a7ba-7551e3a7a968 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.322384] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcebdd8d-7f1c-4255-8cab-d15dc045a2ed {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.327971] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.396988] env[62692]: DEBUG nova.compute.manager [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 800.502560] env[62692]: DEBUG nova.compute.manager [req-d7e58200-9dbe-485f-b12a-a2288905af42 req-2d056e66-224b-441f-aabe-ce85baa2e6f1 service nova] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Received event network-changed-354f1e83-aa2d-441f-8461-937a689e16c2 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 800.502560] env[62692]: DEBUG nova.compute.manager [req-d7e58200-9dbe-485f-b12a-a2288905af42 req-2d056e66-224b-441f-aabe-ce85baa2e6f1 service nova] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Refreshing instance network info cache due to event network-changed-354f1e83-aa2d-441f-8461-937a689e16c2. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 800.502684] env[62692]: DEBUG oslo_concurrency.lockutils [req-d7e58200-9dbe-485f-b12a-a2288905af42 req-2d056e66-224b-441f-aabe-ce85baa2e6f1 service nova] Acquiring lock "refresh_cache-abee88ff-5cf5-4bf1-91e0-93b19cf30046" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.502803] env[62692]: DEBUG oslo_concurrency.lockutils [req-d7e58200-9dbe-485f-b12a-a2288905af42 req-2d056e66-224b-441f-aabe-ce85baa2e6f1 service nova] Acquired lock "refresh_cache-abee88ff-5cf5-4bf1-91e0-93b19cf30046" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.503990] env[62692]: DEBUG nova.network.neutron [req-d7e58200-9dbe-485f-b12a-a2288905af42 req-2d056e66-224b-441f-aabe-ce85baa2e6f1 service nova] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Refreshing network info cache for port 354f1e83-aa2d-441f-8461-937a689e16c2 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 800.595882] env[62692]: DEBUG nova.compute.manager [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 800.714722] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Task: {'id': task-1140973, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.798376] env[62692]: DEBUG nova.compute.manager [req-0c3bde9d-58d9-4af8-b380-bcc455ed192c req-da1ed376-2766-444f-8151-bf454864ee1f service nova] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Received event network-changed-c77af962-553d-4504-825e-78a6689aee79 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 800.798376] env[62692]: DEBUG nova.compute.manager [req-0c3bde9d-58d9-4af8-b380-bcc455ed192c req-da1ed376-2766-444f-8151-bf454864ee1f service nova] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Refreshing instance network info cache due to event network-changed-c77af962-553d-4504-825e-78a6689aee79. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 800.798376] env[62692]: DEBUG oslo_concurrency.lockutils [req-0c3bde9d-58d9-4af8-b380-bcc455ed192c req-da1ed376-2766-444f-8151-bf454864ee1f service nova] Acquiring lock "refresh_cache-bb547773-d176-4c8e-a0fa-a374d5050b1b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.819022] env[62692]: DEBUG oslo_concurrency.lockutils [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.617s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.819022] env[62692]: DEBUG nova.compute.manager [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 800.821545] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Releasing lock "refresh_cache-bb547773-d176-4c8e-a0fa-a374d5050b1b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.822155] env[62692]: DEBUG nova.compute.manager [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Instance network_info: |[{"id": "c77af962-553d-4504-825e-78a6689aee79", "address": "fa:16:3e:9a:01:cc", "network": {"id": "17e72c3b-603f-4305-96ac-89a9b4c9c7e9", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-123501993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7db5465cb5db465b92dfa570572b6e80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc77af962-55", "ovs_interfaceid": "c77af962-553d-4504-825e-78a6689aee79", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 800.829419] env[62692]: DEBUG nova.virt.hardware [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 800.831262] env[62692]: DEBUG nova.virt.hardware [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 800.832389] env[62692]: DEBUG nova.virt.hardware [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 800.832769] env[62692]: DEBUG nova.virt.hardware [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 800.833075] env[62692]: DEBUG nova.virt.hardware [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 800.834020] env[62692]: DEBUG nova.virt.hardware [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 800.834020] env[62692]: DEBUG nova.virt.hardware [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 800.834020] env[62692]: DEBUG nova.virt.hardware [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 800.834020] env[62692]: DEBUG nova.virt.hardware [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 800.834323] env[62692]: DEBUG nova.virt.hardware [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 800.834619] env[62692]: DEBUG nova.virt.hardware [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 800.835335] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.487s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.837090] env[62692]: INFO nova.compute.claims [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 800.839948] env[62692]: DEBUG oslo_concurrency.lockutils [req-0c3bde9d-58d9-4af8-b380-bcc455ed192c req-da1ed376-2766-444f-8151-bf454864ee1f service nova] Acquired lock "refresh_cache-bb547773-d176-4c8e-a0fa-a374d5050b1b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.840612] env[62692]: DEBUG nova.network.neutron [req-0c3bde9d-58d9-4af8-b380-bcc455ed192c req-da1ed376-2766-444f-8151-bf454864ee1f service nova] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Refreshing network info cache for port c77af962-553d-4504-825e-78a6689aee79 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 800.845030] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9a:01:cc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '12d8eedb-97cb-4d3b-b364-42d7fd8b3c85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c77af962-553d-4504-825e-78a6689aee79', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 800.850710] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Creating folder: Project (7db5465cb5db465b92dfa570572b6e80). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 800.852895] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6920a4e8-5bdc-49eb-a704-c01c9b69acd2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.862539] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-06213cbd-ff71-43a9-aa63-660f749c2edb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.865930] env[62692]: INFO nova.compute.manager [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Took 43.96 seconds to build instance. [ 800.875638] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd7fc318-281e-4c63-ab1e-8d410e44242f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.885132] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Created folder: Project (7db5465cb5db465b92dfa570572b6e80) in parent group-v248868. [ 800.885357] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Creating folder: Instances. Parent ref: group-v248902. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 800.886985] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b0734ec8-1886-4c8f-b96b-98b52409537c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.899478] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5a:54:50', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c842425c-544e-4ce2-9657-512723bd318e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f19989c4-0ce9-47ae-989b-0faab24bc9b1', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 800.909351] env[62692]: DEBUG oslo.service.loopingcall [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 800.914738] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 800.915400] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Created folder: Instances in parent group-v248902. [ 800.915635] env[62692]: DEBUG oslo.service.loopingcall [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 800.918379] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-28de4bf4-dce0-4d49-b6e5-4075792e5e8f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.933776] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 800.936253] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3665041c-93d7-49af-8929-4885faba915b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.953766] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.959496] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 800.959496] env[62692]: value = "task-1140977" [ 800.959496] env[62692]: _type = "Task" [ 800.959496] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.961678] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 800.961678] env[62692]: value = "task-1140976" [ 800.961678] env[62692]: _type = "Task" [ 800.961678] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.973329] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140977, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.976647] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140976, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.134614] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.214424] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Task: {'id': task-1140973, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.84451} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.217481] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 105e6fd5-4eff-4a0c-9a4b-e6deade781d5/105e6fd5-4eff-4a0c-9a4b-e6deade781d5.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 801.217481] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 801.217481] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-760bbca1-cdba-4564-aa13-18eb29674632 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.225546] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Waiting for the task: (returnval){ [ 801.225546] env[62692]: value = "task-1140978" [ 801.225546] env[62692]: _type = "Task" [ 801.225546] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.238024] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Task: {'id': task-1140978, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.291355] env[62692]: DEBUG nova.network.neutron [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Successfully updated port: 920fbe7b-7ba1-4d33-bf4f-d96609ed6640 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 801.336797] env[62692]: DEBUG nova.compute.utils [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 801.338271] env[62692]: DEBUG nova.compute.manager [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 801.338436] env[62692]: DEBUG nova.network.neutron [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 801.370226] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a06577b0-590c-4e6e-aab2-5ef07534729c tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "9b8e830c-61b7-4dd2-8324-d3a96eec1465" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 142.686s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.447902] env[62692]: DEBUG nova.policy [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3071751b670a49b889fff6325923df34', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '156dc55f4b764e9289e3003453fe4f27', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 801.469421] env[62692]: DEBUG nova.network.neutron [req-d7e58200-9dbe-485f-b12a-a2288905af42 req-2d056e66-224b-441f-aabe-ce85baa2e6f1 service nova] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Updated VIF entry in instance network info cache for port 354f1e83-aa2d-441f-8461-937a689e16c2. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 801.471079] env[62692]: DEBUG nova.network.neutron [req-d7e58200-9dbe-485f-b12a-a2288905af42 req-2d056e66-224b-441f-aabe-ce85baa2e6f1 service nova] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Updating instance_info_cache with network_info: [{"id": "354f1e83-aa2d-441f-8461-937a689e16c2", "address": "fa:16:3e:7a:37:15", "network": {"id": "14d97ab8-1b79-4ef2-a354-6398d00bdeba", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-707295669-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fc06bb980c584d2980ee5eb48e678772", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap354f1e83-aa", "ovs_interfaceid": "354f1e83-aa2d-441f-8461-937a689e16c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.482592] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140977, 'name': CreateVM_Task, 'duration_secs': 0.504384} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.489840] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 801.489840] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140976, 'name': CreateVM_Task, 'duration_secs': 0.490704} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.490778] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.490942] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.491579] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 801.491845] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 801.492387] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f2e7cd3-bca5-46fb-aa8a-8ff2aaf288f9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.494658] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.494940] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.495238] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 801.496280] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-610edaed-4be7-4344-abf8-d53eb767e562 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.503293] env[62692]: DEBUG oslo_vmware.api [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 801.503293] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]527da735-8f5f-c864-db6e-ea5fee702e0b" [ 801.503293] env[62692]: _type = "Task" [ 801.503293] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.504693] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 801.504693] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]522394b0-67a6-bcac-01d9-bc4b9552d073" [ 801.504693] env[62692]: _type = "Task" [ 801.504693] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.520243] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]522394b0-67a6-bcac-01d9-bc4b9552d073, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.521856] env[62692]: DEBUG oslo_vmware.api [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]527da735-8f5f-c864-db6e-ea5fee702e0b, 'name': SearchDatastore_Task} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.522342] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.522660] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 801.522894] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.523050] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.523228] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 801.523512] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0ed9ceb3-0f3e-41f9-9fb9-d90d99cb2a48 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.534556] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 801.534556] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 801.539043] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aebddd55-f8a8-4cf1-bf02-9c3158148f80 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.544528] env[62692]: DEBUG oslo_vmware.api [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 801.544528] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52fc0cf4-7f65-8a4a-3ff1-463cf4572865" [ 801.544528] env[62692]: _type = "Task" [ 801.544528] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.554357] env[62692]: DEBUG oslo_vmware.api [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52fc0cf4-7f65-8a4a-3ff1-463cf4572865, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.737082] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Task: {'id': task-1140978, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079897} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.737407] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 801.738763] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-621220ee-8682-4938-a524-e514ba641542 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.767515] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] 105e6fd5-4eff-4a0c-9a4b-e6deade781d5/105e6fd5-4eff-4a0c-9a4b-e6deade781d5.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 801.771296] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ab57628a-2f1c-4046-8d40-d250cde7c7b1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.796792] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "refresh_cache-ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.797331] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquired lock "refresh_cache-ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.797331] env[62692]: DEBUG nova.network.neutron [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 801.804990] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Waiting for the task: (returnval){ [ 801.804990] env[62692]: value = "task-1140980" [ 801.804990] env[62692]: _type = "Task" [ 801.804990] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.817700] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Task: {'id': task-1140980, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.843048] env[62692]: DEBUG nova.compute.manager [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 801.872554] env[62692]: DEBUG nova.compute.manager [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 801.880033] env[62692]: DEBUG nova.network.neutron [req-0c3bde9d-58d9-4af8-b380-bcc455ed192c req-da1ed376-2766-444f-8151-bf454864ee1f service nova] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Updated VIF entry in instance network info cache for port c77af962-553d-4504-825e-78a6689aee79. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 801.880033] env[62692]: DEBUG nova.network.neutron [req-0c3bde9d-58d9-4af8-b380-bcc455ed192c req-da1ed376-2766-444f-8151-bf454864ee1f service nova] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Updating instance_info_cache with network_info: [{"id": "c77af962-553d-4504-825e-78a6689aee79", "address": "fa:16:3e:9a:01:cc", "network": {"id": "17e72c3b-603f-4305-96ac-89a9b4c9c7e9", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-123501993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7db5465cb5db465b92dfa570572b6e80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc77af962-55", "ovs_interfaceid": "c77af962-553d-4504-825e-78a6689aee79", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.979631] env[62692]: DEBUG oslo_concurrency.lockutils [req-d7e58200-9dbe-485f-b12a-a2288905af42 req-2d056e66-224b-441f-aabe-ce85baa2e6f1 service nova] Releasing lock "refresh_cache-abee88ff-5cf5-4bf1-91e0-93b19cf30046" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.031021] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]522394b0-67a6-bcac-01d9-bc4b9552d073, 'name': SearchDatastore_Task, 'duration_secs': 0.014014} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.031021] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.031021] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 802.031021] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.031478] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.031478] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 802.031478] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-69e81115-7123-4eff-8906-d09b7eb49385 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.041266] env[62692]: DEBUG nova.network.neutron [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Successfully created port: e047ca79-8094-49de-809c-036fee17d2c5 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 802.046786] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 802.047321] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 802.050960] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45d14589-785a-487d-bf19-a76435555f41 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.060301] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 802.060301] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]523e39db-ccb0-8fa9-b2fa-3459dee2f25c" [ 802.060301] env[62692]: _type = "Task" [ 802.060301] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.064852] env[62692]: DEBUG oslo_vmware.api [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52fc0cf4-7f65-8a4a-3ff1-463cf4572865, 'name': SearchDatastore_Task, 'duration_secs': 0.013182} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.071321] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d8be4d37-2513-4450-9e1b-01d0e64f3946 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.080767] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]523e39db-ccb0-8fa9-b2fa-3459dee2f25c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.082331] env[62692]: DEBUG oslo_vmware.api [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 802.082331] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5276e30b-5ed6-bcf9-03dd-c61eb958fa82" [ 802.082331] env[62692]: _type = "Task" [ 802.082331] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.091296] env[62692]: DEBUG oslo_vmware.api [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5276e30b-5ed6-bcf9-03dd-c61eb958fa82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.320143] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Task: {'id': task-1140980, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.336562] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5238a939-ddbb-4c28-a816-f75edb8b335d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.346186] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-178140b0-1964-4774-bafb-71ab5aeec407 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.388259] env[62692]: DEBUG oslo_concurrency.lockutils [req-0c3bde9d-58d9-4af8-b380-bcc455ed192c req-da1ed376-2766-444f-8151-bf454864ee1f service nova] Releasing lock "refresh_cache-bb547773-d176-4c8e-a0fa-a374d5050b1b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.392615] env[62692]: DEBUG nova.network.neutron [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 802.395222] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5270991-04ee-472f-bb82-cdacc85a38f2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.405587] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a7848e8-54b2-4d49-a773-4ffcc9432257 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.421440] env[62692]: DEBUG nova.compute.provider_tree [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 802.423227] env[62692]: DEBUG oslo_concurrency.lockutils [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.577667] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]523e39db-ccb0-8fa9-b2fa-3459dee2f25c, 'name': SearchDatastore_Task, 'duration_secs': 0.024646} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.578555] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a78fdc1a-05c8-413c-8d95-9db6dd406862 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.585723] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 802.585723] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]528bac9b-83e0-f401-59f2-d0a72187c8d4" [ 802.585723] env[62692]: _type = "Task" [ 802.585723] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.603857] env[62692]: DEBUG oslo_vmware.api [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5276e30b-5ed6-bcf9-03dd-c61eb958fa82, 'name': SearchDatastore_Task, 'duration_secs': 0.029864} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.605445] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.605916] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] bb547773-d176-4c8e-a0fa-a374d5050b1b/bb547773-d176-4c8e-a0fa-a374d5050b1b.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 802.606038] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]528bac9b-83e0-f401-59f2-d0a72187c8d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.606280] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-556875ac-fdb4-494e-9047-94b0f4ba9797 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.616092] env[62692]: DEBUG oslo_vmware.api [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 802.616092] env[62692]: value = "task-1140981" [ 802.616092] env[62692]: _type = "Task" [ 802.616092] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.628983] env[62692]: DEBUG oslo_vmware.api [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1140981, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.656552] env[62692]: DEBUG nova.network.neutron [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Updating instance_info_cache with network_info: [{"id": "920fbe7b-7ba1-4d33-bf4f-d96609ed6640", "address": "fa:16:3e:5e:35:e2", "network": {"id": "17e72c3b-603f-4305-96ac-89a9b4c9c7e9", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-123501993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7db5465cb5db465b92dfa570572b6e80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap920fbe7b-7b", "ovs_interfaceid": "920fbe7b-7ba1-4d33-bf4f-d96609ed6640", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.757515] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "754b3c65-1e4b-49d2-8980-095d975edb01" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.757785] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "754b3c65-1e4b-49d2-8980-095d975edb01" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.817941] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Task: {'id': task-1140980, 'name': ReconfigVM_Task, 'duration_secs': 0.851677} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.817941] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Reconfigured VM instance instance-00000028 to attach disk [datastore2] 105e6fd5-4eff-4a0c-9a4b-e6deade781d5/105e6fd5-4eff-4a0c-9a4b-e6deade781d5.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 802.818217] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aac40bce-8db2-4201-9ba3-31c6a00b88f4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.827661] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "26c1ff5b-9eda-4131-a4ea-d4511a2364b7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.827770] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "26c1ff5b-9eda-4131-a4ea-d4511a2364b7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.828020] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Waiting for the task: (returnval){ [ 802.828020] env[62692]: value = "task-1140982" [ 802.828020] env[62692]: _type = "Task" [ 802.828020] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.836294] env[62692]: DEBUG nova.compute.manager [req-90030974-9f21-4fc7-a3e4-c9a78a513231 req-f0589d2e-300f-49a6-8f9b-ed7b206dedbd service nova] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Received event network-vif-plugged-920fbe7b-7ba1-4d33-bf4f-d96609ed6640 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 802.836577] env[62692]: DEBUG oslo_concurrency.lockutils [req-90030974-9f21-4fc7-a3e4-c9a78a513231 req-f0589d2e-300f-49a6-8f9b-ed7b206dedbd service nova] Acquiring lock "ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.836810] env[62692]: DEBUG oslo_concurrency.lockutils [req-90030974-9f21-4fc7-a3e4-c9a78a513231 req-f0589d2e-300f-49a6-8f9b-ed7b206dedbd service nova] Lock "ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.836889] env[62692]: DEBUG oslo_concurrency.lockutils [req-90030974-9f21-4fc7-a3e4-c9a78a513231 req-f0589d2e-300f-49a6-8f9b-ed7b206dedbd service nova] Lock "ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.837022] env[62692]: DEBUG nova.compute.manager [req-90030974-9f21-4fc7-a3e4-c9a78a513231 req-f0589d2e-300f-49a6-8f9b-ed7b206dedbd service nova] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] No waiting events found dispatching network-vif-plugged-920fbe7b-7ba1-4d33-bf4f-d96609ed6640 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 802.837180] env[62692]: WARNING nova.compute.manager [req-90030974-9f21-4fc7-a3e4-c9a78a513231 req-f0589d2e-300f-49a6-8f9b-ed7b206dedbd service nova] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Received unexpected event network-vif-plugged-920fbe7b-7ba1-4d33-bf4f-d96609ed6640 for instance with vm_state building and task_state spawning. [ 802.837333] env[62692]: DEBUG nova.compute.manager [req-90030974-9f21-4fc7-a3e4-c9a78a513231 req-f0589d2e-300f-49a6-8f9b-ed7b206dedbd service nova] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Received event network-changed-920fbe7b-7ba1-4d33-bf4f-d96609ed6640 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 802.837478] env[62692]: DEBUG nova.compute.manager [req-90030974-9f21-4fc7-a3e4-c9a78a513231 req-f0589d2e-300f-49a6-8f9b-ed7b206dedbd service nova] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Refreshing instance network info cache due to event network-changed-920fbe7b-7ba1-4d33-bf4f-d96609ed6640. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 802.837646] env[62692]: DEBUG oslo_concurrency.lockutils [req-90030974-9f21-4fc7-a3e4-c9a78a513231 req-f0589d2e-300f-49a6-8f9b-ed7b206dedbd service nova] Acquiring lock "refresh_cache-ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.844343] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Task: {'id': task-1140982, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.857066] env[62692]: DEBUG nova.compute.manager [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 802.886910] env[62692]: DEBUG nova.virt.hardware [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 802.887185] env[62692]: DEBUG nova.virt.hardware [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 802.887349] env[62692]: DEBUG nova.virt.hardware [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 802.887557] env[62692]: DEBUG nova.virt.hardware [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 802.887694] env[62692]: DEBUG nova.virt.hardware [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 802.887838] env[62692]: DEBUG nova.virt.hardware [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 802.888077] env[62692]: DEBUG nova.virt.hardware [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 802.888244] env[62692]: DEBUG nova.virt.hardware [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 802.888423] env[62692]: DEBUG nova.virt.hardware [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 802.888637] env[62692]: DEBUG nova.virt.hardware [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 802.888825] env[62692]: DEBUG nova.virt.hardware [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 802.889839] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cf35b07-f8ab-4dd8-8c8a-b458fdb9a55c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.902368] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48c73c81-32b5-4dc0-a544-b0428fe9f6fb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.924836] env[62692]: DEBUG nova.scheduler.client.report [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 803.101376] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]528bac9b-83e0-f401-59f2-d0a72187c8d4, 'name': SearchDatastore_Task, 'duration_secs': 0.018458} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.101687] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.101901] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] fa7bb883-0065-4ecc-9eb5-8c5443344a2a/fa7bb883-0065-4ecc-9eb5-8c5443344a2a.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 803.102235] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-96aa9d0e-c162-4b17-971c-4648f401bba9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.111443] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 803.111443] env[62692]: value = "task-1140983" [ 803.111443] env[62692]: _type = "Task" [ 803.111443] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.126143] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140983, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.129734] env[62692]: DEBUG oslo_vmware.api [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1140981, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.159810] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Releasing lock "refresh_cache-ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.160294] env[62692]: DEBUG nova.compute.manager [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Instance network_info: |[{"id": "920fbe7b-7ba1-4d33-bf4f-d96609ed6640", "address": "fa:16:3e:5e:35:e2", "network": {"id": "17e72c3b-603f-4305-96ac-89a9b4c9c7e9", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-123501993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7db5465cb5db465b92dfa570572b6e80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap920fbe7b-7b", "ovs_interfaceid": "920fbe7b-7ba1-4d33-bf4f-d96609ed6640", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 803.160669] env[62692]: DEBUG oslo_concurrency.lockutils [req-90030974-9f21-4fc7-a3e4-c9a78a513231 req-f0589d2e-300f-49a6-8f9b-ed7b206dedbd service nova] Acquired lock "refresh_cache-ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.160894] env[62692]: DEBUG nova.network.neutron [req-90030974-9f21-4fc7-a3e4-c9a78a513231 req-f0589d2e-300f-49a6-8f9b-ed7b206dedbd service nova] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Refreshing network info cache for port 920fbe7b-7ba1-4d33-bf4f-d96609ed6640 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 803.162449] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5e:35:e2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '12d8eedb-97cb-4d3b-b364-42d7fd8b3c85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '920fbe7b-7ba1-4d33-bf4f-d96609ed6640', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 803.170435] env[62692]: DEBUG oslo.service.loopingcall [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 803.171535] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 803.171770] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-03aec14b-93d7-4c27-8bcc-8f7688fa42ba {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.197143] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 803.197143] env[62692]: value = "task-1140984" [ 803.197143] env[62692]: _type = "Task" [ 803.197143] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.207152] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140984, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.340163] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Task: {'id': task-1140982, 'name': Rename_Task, 'duration_secs': 0.221229} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.340509] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 803.340777] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ade25156-1559-4fe5-86dc-0706d988bbf1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.348257] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Waiting for the task: (returnval){ [ 803.348257] env[62692]: value = "task-1140985" [ 803.348257] env[62692]: _type = "Task" [ 803.348257] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.356375] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Task: {'id': task-1140985, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.430572] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.595s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.431400] env[62692]: DEBUG nova.compute.manager [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 803.434802] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.903s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.436345] env[62692]: INFO nova.compute.claims [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 803.629765] env[62692]: DEBUG oslo_vmware.api [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1140981, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.704711} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.633694] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] bb547773-d176-4c8e-a0fa-a374d5050b1b/bb547773-d176-4c8e-a0fa-a374d5050b1b.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 803.633949] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 803.634303] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140983, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.634523] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d3530414-e2f6-4cda-abc6-c03790b9dd58 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.644777] env[62692]: DEBUG oslo_vmware.api [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 803.644777] env[62692]: value = "task-1140986" [ 803.644777] env[62692]: _type = "Task" [ 803.644777] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.656255] env[62692]: DEBUG oslo_vmware.api [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1140986, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.713137] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140984, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.871131] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Task: {'id': task-1140985, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.940788] env[62692]: DEBUG nova.compute.utils [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 803.946141] env[62692]: DEBUG nova.compute.manager [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 803.946141] env[62692]: DEBUG nova.network.neutron [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 804.015103] env[62692]: DEBUG nova.policy [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a251ef5f6cb143178928babc75a0fef0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ba97a775e99a463c845518b0df8cd3e2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 804.128491] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140983, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.893571} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.128820] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] fa7bb883-0065-4ecc-9eb5-8c5443344a2a/fa7bb883-0065-4ecc-9eb5-8c5443344a2a.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 804.129034] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 804.129307] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-954a0337-8b17-4c92-969d-fbe43f3e04cd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.137419] env[62692]: DEBUG nova.network.neutron [req-90030974-9f21-4fc7-a3e4-c9a78a513231 req-f0589d2e-300f-49a6-8f9b-ed7b206dedbd service nova] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Updated VIF entry in instance network info cache for port 920fbe7b-7ba1-4d33-bf4f-d96609ed6640. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 804.137769] env[62692]: DEBUG nova.network.neutron [req-90030974-9f21-4fc7-a3e4-c9a78a513231 req-f0589d2e-300f-49a6-8f9b-ed7b206dedbd service nova] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Updating instance_info_cache with network_info: [{"id": "920fbe7b-7ba1-4d33-bf4f-d96609ed6640", "address": "fa:16:3e:5e:35:e2", "network": {"id": "17e72c3b-603f-4305-96ac-89a9b4c9c7e9", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-123501993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7db5465cb5db465b92dfa570572b6e80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap920fbe7b-7b", "ovs_interfaceid": "920fbe7b-7ba1-4d33-bf4f-d96609ed6640", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.141727] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 804.141727] env[62692]: value = "task-1140987" [ 804.141727] env[62692]: _type = "Task" [ 804.141727] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.158280] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140987, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.160168] env[62692]: DEBUG nova.network.neutron [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Successfully updated port: e047ca79-8094-49de-809c-036fee17d2c5 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 804.166801] env[62692]: DEBUG oslo_vmware.api [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1140986, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.109371} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.167210] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 804.167788] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-087af8b6-a16a-4f8f-960b-287b83191170 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.193669] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Reconfiguring VM instance instance-00000029 to attach disk [datastore2] bb547773-d176-4c8e-a0fa-a374d5050b1b/bb547773-d176-4c8e-a0fa-a374d5050b1b.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 804.194382] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fad75400-1b47-4b0a-82f0-cb2368baed84 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.221451] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140984, 'name': CreateVM_Task, 'duration_secs': 0.664876} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.222876] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 804.223359] env[62692]: DEBUG oslo_vmware.api [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 804.223359] env[62692]: value = "task-1140988" [ 804.223359] env[62692]: _type = "Task" [ 804.223359] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.224208] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.224421] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.225391] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 804.225391] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16ce02d2-974a-421e-a63f-85a258654bbe {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.236241] env[62692]: DEBUG oslo_vmware.api [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1140988, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.238074] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 804.238074] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5275f378-068d-2bd4-1b58-0ec26c2ebbd2" [ 804.238074] env[62692]: _type = "Task" [ 804.238074] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.247407] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5275f378-068d-2bd4-1b58-0ec26c2ebbd2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.361752] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Task: {'id': task-1140985, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.443915] env[62692]: DEBUG nova.compute.manager [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 804.523181] env[62692]: DEBUG nova.network.neutron [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Successfully created port: d4991635-5c62-4473-9661-97ac75a187fb {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 804.642328] env[62692]: DEBUG oslo_concurrency.lockutils [req-90030974-9f21-4fc7-a3e4-c9a78a513231 req-f0589d2e-300f-49a6-8f9b-ed7b206dedbd service nova] Releasing lock "refresh_cache-ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.653859] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140987, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088949} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.656805] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 804.661944] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d3302a-bcc7-452c-a99d-e1e0eff470ce {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.664851] env[62692]: DEBUG oslo_concurrency.lockutils [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Acquiring lock "refresh_cache-6097d6b2-52d5-4765-94c0-d5f3609464d8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.664991] env[62692]: DEBUG oslo_concurrency.lockutils [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Acquired lock "refresh_cache-6097d6b2-52d5-4765-94c0-d5f3609464d8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.665158] env[62692]: DEBUG nova.network.neutron [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 804.692271] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Reconfiguring VM instance instance-00000024 to attach disk [datastore1] fa7bb883-0065-4ecc-9eb5-8c5443344a2a/fa7bb883-0065-4ecc-9eb5-8c5443344a2a.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 804.696770] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-98960697-7048-44d9-9330-14b6909399bb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.719313] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 804.719313] env[62692]: value = "task-1140989" [ 804.719313] env[62692]: _type = "Task" [ 804.719313] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.734595] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140989, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.742502] env[62692]: DEBUG oslo_vmware.api [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1140988, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.757846] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5275f378-068d-2bd4-1b58-0ec26c2ebbd2, 'name': SearchDatastore_Task, 'duration_secs': 0.024341} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.758910] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.759213] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 804.759482] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.760188] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.760188] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 804.760188] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-64cedb09-0985-4f39-a7eb-43b2fda273d9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.771364] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 804.772505] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 804.772505] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1fab9da2-ff85-48fe-ad8c-cb7f0d86ff5c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.777964] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 804.777964] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]526935cc-0555-5b04-2e06-eaa0ea7179b6" [ 804.777964] env[62692]: _type = "Task" [ 804.777964] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.788210] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]526935cc-0555-5b04-2e06-eaa0ea7179b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.859325] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Task: {'id': task-1140985, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.931239] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a874b45-91b2-4bdb-9796-f1457403a681 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.940221] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c5b4493-b1ac-493f-bd4e-644275b5b93a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.945692] env[62692]: DEBUG nova.compute.manager [req-69d9a562-8ca9-48e2-8f0d-b55e3b91a707 req-043c4025-2de4-498e-a6d0-67db001616e3 service nova] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Received event network-vif-plugged-e047ca79-8094-49de-809c-036fee17d2c5 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 804.945916] env[62692]: DEBUG oslo_concurrency.lockutils [req-69d9a562-8ca9-48e2-8f0d-b55e3b91a707 req-043c4025-2de4-498e-a6d0-67db001616e3 service nova] Acquiring lock "6097d6b2-52d5-4765-94c0-d5f3609464d8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.946159] env[62692]: DEBUG oslo_concurrency.lockutils [req-69d9a562-8ca9-48e2-8f0d-b55e3b91a707 req-043c4025-2de4-498e-a6d0-67db001616e3 service nova] Lock "6097d6b2-52d5-4765-94c0-d5f3609464d8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.946299] env[62692]: DEBUG oslo_concurrency.lockutils [req-69d9a562-8ca9-48e2-8f0d-b55e3b91a707 req-043c4025-2de4-498e-a6d0-67db001616e3 service nova] Lock "6097d6b2-52d5-4765-94c0-d5f3609464d8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.946467] env[62692]: DEBUG nova.compute.manager [req-69d9a562-8ca9-48e2-8f0d-b55e3b91a707 req-043c4025-2de4-498e-a6d0-67db001616e3 service nova] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] No waiting events found dispatching network-vif-plugged-e047ca79-8094-49de-809c-036fee17d2c5 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 804.946676] env[62692]: WARNING nova.compute.manager [req-69d9a562-8ca9-48e2-8f0d-b55e3b91a707 req-043c4025-2de4-498e-a6d0-67db001616e3 service nova] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Received unexpected event network-vif-plugged-e047ca79-8094-49de-809c-036fee17d2c5 for instance with vm_state building and task_state spawning. [ 804.946883] env[62692]: DEBUG nova.compute.manager [req-69d9a562-8ca9-48e2-8f0d-b55e3b91a707 req-043c4025-2de4-498e-a6d0-67db001616e3 service nova] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Received event network-changed-e047ca79-8094-49de-809c-036fee17d2c5 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 804.947091] env[62692]: DEBUG nova.compute.manager [req-69d9a562-8ca9-48e2-8f0d-b55e3b91a707 req-043c4025-2de4-498e-a6d0-67db001616e3 service nova] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Refreshing instance network info cache due to event network-changed-e047ca79-8094-49de-809c-036fee17d2c5. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 804.947270] env[62692]: DEBUG oslo_concurrency.lockutils [req-69d9a562-8ca9-48e2-8f0d-b55e3b91a707 req-043c4025-2de4-498e-a6d0-67db001616e3 service nova] Acquiring lock "refresh_cache-6097d6b2-52d5-4765-94c0-d5f3609464d8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.981567] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9a34178-4b30-40c2-bc0c-ed0995198328 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.991501] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd478712-b29a-45af-a7ae-79610c7add26 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.007585] env[62692]: DEBUG nova.compute.provider_tree [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 805.198347] env[62692]: DEBUG nova.network.neutron [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 805.233541] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140989, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.239696] env[62692]: DEBUG oslo_vmware.api [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1140988, 'name': ReconfigVM_Task, 'duration_secs': 0.918817} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.242284] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Reconfigured VM instance instance-00000029 to attach disk [datastore2] bb547773-d176-4c8e-a0fa-a374d5050b1b/bb547773-d176-4c8e-a0fa-a374d5050b1b.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 805.242965] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d45ff9be-1f66-4338-a5d4-a7b308a3ab03 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.251172] env[62692]: DEBUG oslo_vmware.api [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 805.251172] env[62692]: value = "task-1140990" [ 805.251172] env[62692]: _type = "Task" [ 805.251172] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.260589] env[62692]: DEBUG oslo_vmware.api [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1140990, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.295561] env[62692]: DEBUG oslo_concurrency.lockutils [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Acquiring lock "f44e2ee2-6321-4237-9644-a19b5e90ae16" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.295814] env[62692]: DEBUG oslo_concurrency.lockutils [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Lock "f44e2ee2-6321-4237-9644-a19b5e90ae16" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.296045] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]526935cc-0555-5b04-2e06-eaa0ea7179b6, 'name': SearchDatastore_Task, 'duration_secs': 0.025232} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.296831] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-187cde31-5c00-462c-88e9-216839496c85 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.303416] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 805.303416] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52affe7e-8b9a-d983-06bd-1c5275aa91f1" [ 805.303416] env[62692]: _type = "Task" [ 805.303416] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.313121] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52affe7e-8b9a-d983-06bd-1c5275aa91f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.360819] env[62692]: DEBUG oslo_vmware.api [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Task: {'id': task-1140985, 'name': PowerOnVM_Task, 'duration_secs': 1.982906} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.361273] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 805.361356] env[62692]: INFO nova.compute.manager [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Took 10.35 seconds to spawn the instance on the hypervisor. [ 805.361530] env[62692]: DEBUG nova.compute.manager [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 805.362299] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfd143d2-0629-44e4-84b0-c425882e29fa {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.376266] env[62692]: DEBUG nova.network.neutron [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Updating instance_info_cache with network_info: [{"id": "e047ca79-8094-49de-809c-036fee17d2c5", "address": "fa:16:3e:bc:00:68", "network": {"id": "4e0237af-dde2-487a-bcb9-4abe51b65c22", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-949976397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "156dc55f4b764e9289e3003453fe4f27", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "309d7cfa-b4da-4eec-9f4b-2e10d215fac7", "external-id": "nsx-vlan-transportzone-285", "segmentation_id": 285, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape047ca79-80", "ovs_interfaceid": "e047ca79-8094-49de-809c-036fee17d2c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.482994] env[62692]: DEBUG nova.compute.manager [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 805.510965] env[62692]: DEBUG nova.scheduler.client.report [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 805.525442] env[62692]: DEBUG nova.virt.hardware [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 805.525714] env[62692]: DEBUG nova.virt.hardware [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 805.526515] env[62692]: DEBUG nova.virt.hardware [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 805.526515] env[62692]: DEBUG nova.virt.hardware [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 805.526515] env[62692]: DEBUG nova.virt.hardware [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 805.526515] env[62692]: DEBUG nova.virt.hardware [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 805.527698] env[62692]: DEBUG nova.virt.hardware [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 805.527698] env[62692]: DEBUG nova.virt.hardware [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 805.527698] env[62692]: DEBUG nova.virt.hardware [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 805.527698] env[62692]: DEBUG nova.virt.hardware [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 805.527698] env[62692]: DEBUG nova.virt.hardware [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 805.528296] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dee2f21e-1fa6-46b0-b0e9-7271e189ce9c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.539946] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fadabfe-83d7-493d-972d-27f40fad2ad8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.730170] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140989, 'name': ReconfigVM_Task, 'duration_secs': 0.830611} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.730501] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Reconfigured VM instance instance-00000024 to attach disk [datastore1] fa7bb883-0065-4ecc-9eb5-8c5443344a2a/fa7bb883-0065-4ecc-9eb5-8c5443344a2a.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 805.731230] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dca61556-e3d6-4864-bb7d-2a8b31b0ce89 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.742084] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 805.742084] env[62692]: value = "task-1140991" [ 805.742084] env[62692]: _type = "Task" [ 805.742084] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.756575] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140991, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.769042] env[62692]: DEBUG oslo_vmware.api [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1140990, 'name': Rename_Task, 'duration_secs': 0.359299} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.769042] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 805.769042] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8145dd4b-ec01-4489-85fb-44746a6398ce {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.776829] env[62692]: DEBUG oslo_vmware.api [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 805.776829] env[62692]: value = "task-1140992" [ 805.776829] env[62692]: _type = "Task" [ 805.776829] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.786539] env[62692]: DEBUG oslo_vmware.api [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1140992, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.814257] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52affe7e-8b9a-d983-06bd-1c5275aa91f1, 'name': SearchDatastore_Task, 'duration_secs': 0.025043} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.814514] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.814788] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64/ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 805.815058] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d4c9ad33-e547-4723-828c-1ee702e38995 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.823609] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 805.823609] env[62692]: value = "task-1140993" [ 805.823609] env[62692]: _type = "Task" [ 805.823609] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.833632] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1140993, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.879992] env[62692]: DEBUG oslo_concurrency.lockutils [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Releasing lock "refresh_cache-6097d6b2-52d5-4765-94c0-d5f3609464d8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.879992] env[62692]: DEBUG nova.compute.manager [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Instance network_info: |[{"id": "e047ca79-8094-49de-809c-036fee17d2c5", "address": "fa:16:3e:bc:00:68", "network": {"id": "4e0237af-dde2-487a-bcb9-4abe51b65c22", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-949976397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "156dc55f4b764e9289e3003453fe4f27", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "309d7cfa-b4da-4eec-9f4b-2e10d215fac7", "external-id": "nsx-vlan-transportzone-285", "segmentation_id": 285, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape047ca79-80", "ovs_interfaceid": "e047ca79-8094-49de-809c-036fee17d2c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 805.880227] env[62692]: INFO nova.compute.manager [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Took 32.68 seconds to build instance. [ 805.881449] env[62692]: DEBUG oslo_concurrency.lockutils [req-69d9a562-8ca9-48e2-8f0d-b55e3b91a707 req-043c4025-2de4-498e-a6d0-67db001616e3 service nova] Acquired lock "refresh_cache-6097d6b2-52d5-4765-94c0-d5f3609464d8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.881776] env[62692]: DEBUG nova.network.neutron [req-69d9a562-8ca9-48e2-8f0d-b55e3b91a707 req-043c4025-2de4-498e-a6d0-67db001616e3 service nova] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Refreshing network info cache for port e047ca79-8094-49de-809c-036fee17d2c5 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 805.883032] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bc:00:68', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '309d7cfa-b4da-4eec-9f4b-2e10d215fac7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e047ca79-8094-49de-809c-036fee17d2c5', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 805.891727] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Creating folder: Project (156dc55f4b764e9289e3003453fe4f27). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 805.895139] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d7d957cc-822e-4365-a49c-deb5d0a21124 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.909189] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Created folder: Project (156dc55f4b764e9289e3003453fe4f27) in parent group-v248868. [ 805.909653] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Creating folder: Instances. Parent ref: group-v248907. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 805.909719] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7853c9fa-60b2-445e-99f0-5681b2709c34 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.921669] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Created folder: Instances in parent group-v248907. [ 805.921967] env[62692]: DEBUG oslo.service.loopingcall [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 805.922208] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 805.922450] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e5421229-5ef1-4f37-8b24-c1ff43afa418 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.944720] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 805.944720] env[62692]: value = "task-1140996" [ 805.944720] env[62692]: _type = "Task" [ 805.944720] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.952803] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140996, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.019816] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.585s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.020511] env[62692]: DEBUG nova.compute.manager [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 806.023734] env[62692]: DEBUG oslo_concurrency.lockutils [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.450s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.025385] env[62692]: INFO nova.compute.claims [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 806.204668] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Acquiring lock "105e6fd5-4eff-4a0c-9a4b-e6deade781d5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.211021] env[62692]: DEBUG nova.network.neutron [req-69d9a562-8ca9-48e2-8f0d-b55e3b91a707 req-043c4025-2de4-498e-a6d0-67db001616e3 service nova] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Updated VIF entry in instance network info cache for port e047ca79-8094-49de-809c-036fee17d2c5. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 806.211617] env[62692]: DEBUG nova.network.neutron [req-69d9a562-8ca9-48e2-8f0d-b55e3b91a707 req-043c4025-2de4-498e-a6d0-67db001616e3 service nova] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Updating instance_info_cache with network_info: [{"id": "e047ca79-8094-49de-809c-036fee17d2c5", "address": "fa:16:3e:bc:00:68", "network": {"id": "4e0237af-dde2-487a-bcb9-4abe51b65c22", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-949976397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "156dc55f4b764e9289e3003453fe4f27", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "309d7cfa-b4da-4eec-9f4b-2e10d215fac7", "external-id": "nsx-vlan-transportzone-285", "segmentation_id": 285, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape047ca79-80", "ovs_interfaceid": "e047ca79-8094-49de-809c-036fee17d2c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.256980] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140991, 'name': Rename_Task, 'duration_secs': 0.198639} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.257290] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 806.257586] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b265aa79-5ab5-450a-9e61-9eb96fbb826b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.270161] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 806.270161] env[62692]: value = "task-1140997" [ 806.270161] env[62692]: _type = "Task" [ 806.270161] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.287497] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140997, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.298941] env[62692]: DEBUG oslo_vmware.api [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1140992, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.311590] env[62692]: DEBUG nova.network.neutron [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Successfully updated port: d4991635-5c62-4473-9661-97ac75a187fb {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 806.327369] env[62692]: DEBUG nova.compute.manager [req-cfc53b62-4450-42c9-95f5-1d9f2fb3b6a0 req-6170863e-647b-4a51-ab13-e78c4466f761 service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] Received event network-vif-plugged-d4991635-5c62-4473-9661-97ac75a187fb {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 806.327588] env[62692]: DEBUG oslo_concurrency.lockutils [req-cfc53b62-4450-42c9-95f5-1d9f2fb3b6a0 req-6170863e-647b-4a51-ab13-e78c4466f761 service nova] Acquiring lock "27583cb4-0478-44d8-8970-478682271513-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.327846] env[62692]: DEBUG oslo_concurrency.lockutils [req-cfc53b62-4450-42c9-95f5-1d9f2fb3b6a0 req-6170863e-647b-4a51-ab13-e78c4466f761 service nova] Lock "27583cb4-0478-44d8-8970-478682271513-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.327964] env[62692]: DEBUG oslo_concurrency.lockutils [req-cfc53b62-4450-42c9-95f5-1d9f2fb3b6a0 req-6170863e-647b-4a51-ab13-e78c4466f761 service nova] Lock "27583cb4-0478-44d8-8970-478682271513-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.328828] env[62692]: DEBUG nova.compute.manager [req-cfc53b62-4450-42c9-95f5-1d9f2fb3b6a0 req-6170863e-647b-4a51-ab13-e78c4466f761 service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] No waiting events found dispatching network-vif-plugged-d4991635-5c62-4473-9661-97ac75a187fb {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 806.332021] env[62692]: WARNING nova.compute.manager [req-cfc53b62-4450-42c9-95f5-1d9f2fb3b6a0 req-6170863e-647b-4a51-ab13-e78c4466f761 service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] Received unexpected event network-vif-plugged-d4991635-5c62-4473-9661-97ac75a187fb for instance with vm_state building and task_state spawning. [ 806.348826] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1140993, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.396806] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a71a8209-8505-4087-9d5f-f04fbe0f93fb tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Lock "105e6fd5-4eff-4a0c-9a4b-e6deade781d5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 146.818s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.398280] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Lock "105e6fd5-4eff-4a0c-9a4b-e6deade781d5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.194s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.399427] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Acquiring lock "105e6fd5-4eff-4a0c-9a4b-e6deade781d5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.399427] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Lock "105e6fd5-4eff-4a0c-9a4b-e6deade781d5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.399427] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Lock "105e6fd5-4eff-4a0c-9a4b-e6deade781d5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.402962] env[62692]: INFO nova.compute.manager [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Terminating instance [ 806.405212] env[62692]: DEBUG nova.compute.manager [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 806.406265] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 806.406387] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f0296cc-db77-4124-950f-3330771d5032 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.419897] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 806.420223] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fd73c4b7-70e4-47c9-8344-45fa394fac45 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.431104] env[62692]: DEBUG oslo_vmware.api [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Waiting for the task: (returnval){ [ 806.431104] env[62692]: value = "task-1140998" [ 806.431104] env[62692]: _type = "Task" [ 806.431104] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.444200] env[62692]: DEBUG oslo_vmware.api [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Task: {'id': task-1140998, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.455878] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1140996, 'name': CreateVM_Task, 'duration_secs': 0.455217} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.456060] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 806.456792] env[62692]: DEBUG oslo_concurrency.lockutils [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.456956] env[62692]: DEBUG oslo_concurrency.lockutils [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.457337] env[62692]: DEBUG oslo_concurrency.lockutils [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 806.457941] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c6dd588-0223-44de-ab46-f242449403f2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.464230] env[62692]: DEBUG oslo_vmware.api [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Waiting for the task: (returnval){ [ 806.464230] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52cb1947-188d-9241-febf-939c67d831a9" [ 806.464230] env[62692]: _type = "Task" [ 806.464230] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.473977] env[62692]: DEBUG oslo_vmware.api [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52cb1947-188d-9241-febf-939c67d831a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.529660] env[62692]: DEBUG nova.compute.utils [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 806.535245] env[62692]: DEBUG nova.compute.manager [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 806.535245] env[62692]: DEBUG nova.network.neutron [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 806.605908] env[62692]: DEBUG nova.policy [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '24e7aa820d314f77ace38aa3a564d449', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4fed583a514c4043a4b5772c73e21aff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 806.715611] env[62692]: DEBUG oslo_concurrency.lockutils [req-69d9a562-8ca9-48e2-8f0d-b55e3b91a707 req-043c4025-2de4-498e-a6d0-67db001616e3 service nova] Releasing lock "refresh_cache-6097d6b2-52d5-4765-94c0-d5f3609464d8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.784805] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140997, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.791064] env[62692]: DEBUG oslo_vmware.api [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1140992, 'name': PowerOnVM_Task, 'duration_secs': 0.990824} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.791222] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 806.791485] env[62692]: INFO nova.compute.manager [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Took 9.24 seconds to spawn the instance on the hypervisor. [ 806.791677] env[62692]: DEBUG nova.compute.manager [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 806.792476] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-801d1983-76d7-4a7d-a1bc-a791bb4aab26 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.814881] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "refresh_cache-27583cb4-0478-44d8-8970-478682271513" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.814881] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "refresh_cache-27583cb4-0478-44d8-8970-478682271513" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.815123] env[62692]: DEBUG nova.network.neutron [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 806.837018] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1140993, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.644733} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.837018] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64/ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 806.837018] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 806.837253] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8b23f3db-68d7-41c5-a21e-c9875589e79b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.846777] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 806.846777] env[62692]: value = "task-1140999" [ 806.846777] env[62692]: _type = "Task" [ 806.846777] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.856358] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1140999, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.901647] env[62692]: DEBUG nova.compute.manager [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 806.936202] env[62692]: DEBUG nova.network.neutron [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Successfully created port: b0db2db5-00e9-4a3a-a41a-7ed7f4985073 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 806.944744] env[62692]: DEBUG oslo_vmware.api [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Task: {'id': task-1140998, 'name': PowerOffVM_Task, 'duration_secs': 0.22551} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.945192] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 806.945579] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 806.945737] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0af2faca-21b8-450c-8942-57efb2ffbcd2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.977975] env[62692]: DEBUG oslo_vmware.api [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52cb1947-188d-9241-febf-939c67d831a9, 'name': SearchDatastore_Task, 'duration_secs': 0.011243} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.978379] env[62692]: DEBUG oslo_concurrency.lockutils [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.978678] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 806.978934] env[62692]: DEBUG oslo_concurrency.lockutils [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.979093] env[62692]: DEBUG oslo_concurrency.lockutils [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.979330] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 806.979577] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-04d142eb-8841-4789-8f43-dfb89d81e5d6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.989949] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 806.990188] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 806.990837] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb68c01e-262b-4b31-8eed-50d0a34e2eb4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.998651] env[62692]: DEBUG oslo_vmware.api [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Waiting for the task: (returnval){ [ 806.998651] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5214db93-c044-6f2a-ab38-fc1078cceea4" [ 806.998651] env[62692]: _type = "Task" [ 806.998651] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.009464] env[62692]: DEBUG oslo_vmware.api [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5214db93-c044-6f2a-ab38-fc1078cceea4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.014121] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 807.014358] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 807.014533] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Deleting the datastore file [datastore2] 105e6fd5-4eff-4a0c-9a4b-e6deade781d5 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 807.014909] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a3144e41-4fad-403f-88b9-ea53b54262e5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.025464] env[62692]: DEBUG oslo_vmware.api [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Waiting for the task: (returnval){ [ 807.025464] env[62692]: value = "task-1141001" [ 807.025464] env[62692]: _type = "Task" [ 807.025464] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.037411] env[62692]: DEBUG nova.compute.manager [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 807.042735] env[62692]: DEBUG oslo_vmware.api [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Task: {'id': task-1141001, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.294989] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140997, 'name': PowerOnVM_Task} progress is 19%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.321759] env[62692]: INFO nova.compute.manager [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Took 27.27 seconds to build instance. [ 807.326709] env[62692]: DEBUG nova.network.neutron [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Successfully created port: 13349e97-1acd-48d2-a24b-44160a60773e {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 807.362326] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1140999, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.143265} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.362430] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 807.363413] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26a3888e-200d-4128-beb8-c92658b4da9a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.367279] env[62692]: DEBUG nova.network.neutron [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 807.397190] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Reconfiguring VM instance instance-0000002a to attach disk [datastore1] ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64/ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 807.401736] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4b901176-5aa5-4dd4-a201-12e1c01763c2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.427654] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 807.427654] env[62692]: value = "task-1141002" [ 807.427654] env[62692]: _type = "Task" [ 807.427654] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.435833] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.443776] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141002, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.511411] env[62692]: DEBUG oslo_vmware.api [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5214db93-c044-6f2a-ab38-fc1078cceea4, 'name': SearchDatastore_Task, 'duration_secs': 0.038771} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.512209] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0da72d79-1f55-4567-b9d3-3d88339938b0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.522012] env[62692]: DEBUG oslo_vmware.api [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Waiting for the task: (returnval){ [ 807.522012] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]528d9d18-3d40-a68a-bd69-746b35117396" [ 807.522012] env[62692]: _type = "Task" [ 807.522012] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.539116] env[62692]: DEBUG oslo_vmware.api [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]528d9d18-3d40-a68a-bd69-746b35117396, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.547482] env[62692]: DEBUG oslo_vmware.api [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Task: {'id': task-1141001, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.549126] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b5384a4-69b6-466e-b480-1291951e7198 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.559462] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12eaec83-6d24-4afe-a99f-3e960a4ef2ac {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.595883] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e47f57e-e2dc-4d16-bab8-fe73e81cd5f4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.598129] env[62692]: DEBUG nova.network.neutron [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Updating instance_info_cache with network_info: [{"id": "d4991635-5c62-4473-9661-97ac75a187fb", "address": "fa:16:3e:d4:98:43", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4991635-5c", "ovs_interfaceid": "d4991635-5c62-4473-9661-97ac75a187fb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.606888] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-360c97cb-4fe0-4241-b63c-0bdea5ba664a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.625262] env[62692]: DEBUG nova.compute.provider_tree [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 807.742863] env[62692]: DEBUG nova.network.neutron [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Successfully created port: 56f47df0-50f1-43e0-a2d5-3a4ad9d19327 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 807.782611] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140997, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.826954] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d8496297-8a1e-4a51-aeee-04f9b65f8fe0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "bb547773-d176-4c8e-a0fa-a374d5050b1b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.165s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.940076] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141002, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.047219] env[62692]: DEBUG oslo_vmware.api [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Task: {'id': task-1141001, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.667494} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.047492] env[62692]: DEBUG oslo_vmware.api [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]528d9d18-3d40-a68a-bd69-746b35117396, 'name': SearchDatastore_Task, 'duration_secs': 0.032228} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.048800] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 808.048800] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 808.049160] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 808.049449] env[62692]: INFO nova.compute.manager [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Took 1.64 seconds to destroy the instance on the hypervisor. [ 808.049544] env[62692]: DEBUG oslo.service.loopingcall [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 808.049769] env[62692]: DEBUG oslo_concurrency.lockutils [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.050059] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 6097d6b2-52d5-4765-94c0-d5f3609464d8/6097d6b2-52d5-4765-94c0-d5f3609464d8.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 808.050264] env[62692]: DEBUG nova.compute.manager [-] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 808.050401] env[62692]: DEBUG nova.network.neutron [-] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 808.052157] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c58e2272-e7e5-44d5-972f-7794f6243613 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.055819] env[62692]: DEBUG nova.compute.manager [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 808.064284] env[62692]: DEBUG oslo_vmware.api [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Waiting for the task: (returnval){ [ 808.064284] env[62692]: value = "task-1141003" [ 808.064284] env[62692]: _type = "Task" [ 808.064284] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.073740] env[62692]: DEBUG oslo_vmware.api [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Task: {'id': task-1141003, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.093965] env[62692]: DEBUG nova.virt.hardware [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 808.094329] env[62692]: DEBUG nova.virt.hardware [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 808.094399] env[62692]: DEBUG nova.virt.hardware [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 808.094550] env[62692]: DEBUG nova.virt.hardware [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 808.094697] env[62692]: DEBUG nova.virt.hardware [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 808.094985] env[62692]: DEBUG nova.virt.hardware [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 808.095070] env[62692]: DEBUG nova.virt.hardware [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 808.095219] env[62692]: DEBUG nova.virt.hardware [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 808.095391] env[62692]: DEBUG nova.virt.hardware [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 808.095555] env[62692]: DEBUG nova.virt.hardware [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 808.095730] env[62692]: DEBUG nova.virt.hardware [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 808.096634] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc41c6b0-f0a8-4fd1-b9f6-09a6c5e34893 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.100979] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "refresh_cache-27583cb4-0478-44d8-8970-478682271513" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.101133] env[62692]: DEBUG nova.compute.manager [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Instance network_info: |[{"id": "d4991635-5c62-4473-9661-97ac75a187fb", "address": "fa:16:3e:d4:98:43", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4991635-5c", "ovs_interfaceid": "d4991635-5c62-4473-9661-97ac75a187fb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 808.101512] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d4:98:43', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6a31b224-77d7-48a4-af87-312758611b8e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd4991635-5c62-4473-9661-97ac75a187fb', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 808.109333] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Creating folder: Project (ba97a775e99a463c845518b0df8cd3e2). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 808.113794] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fd2a6daf-28dc-4cc5-a564-c2a34c8a6a6e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.117706] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f1fa2b6-6f29-4ff3-a465-c97b3596b9ce {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.134446] env[62692]: DEBUG nova.scheduler.client.report [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 808.140253] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Created folder: Project (ba97a775e99a463c845518b0df8cd3e2) in parent group-v248868. [ 808.140253] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Creating folder: Instances. Parent ref: group-v248910. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 808.140471] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-de5de0f9-27fd-4331-ab23-c362f77226ac {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.152363] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Created folder: Instances in parent group-v248910. [ 808.155023] env[62692]: DEBUG oslo.service.loopingcall [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 808.155023] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27583cb4-0478-44d8-8970-478682271513] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 808.155023] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d160bef1-34f5-4920-bbe7-bae333850ad3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.176378] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 808.176378] env[62692]: value = "task-1141006" [ 808.176378] env[62692]: _type = "Task" [ 808.176378] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.185307] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141006, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.285557] env[62692]: DEBUG oslo_vmware.api [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1140997, 'name': PowerOnVM_Task, 'duration_secs': 1.878637} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.285557] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 808.285557] env[62692]: DEBUG nova.compute.manager [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 808.286865] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52806054-5713-40af-b565-ebbf691137fe {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.330368] env[62692]: DEBUG nova.compute.manager [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 808.385349] env[62692]: DEBUG nova.compute.manager [req-8c50b1f7-4efa-42ba-99f9-e3b3b334678a req-63e90339-b36a-4daa-bc49-6198ebcde63a service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] Received event network-changed-d4991635-5c62-4473-9661-97ac75a187fb {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 808.385349] env[62692]: DEBUG nova.compute.manager [req-8c50b1f7-4efa-42ba-99f9-e3b3b334678a req-63e90339-b36a-4daa-bc49-6198ebcde63a service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] Refreshing instance network info cache due to event network-changed-d4991635-5c62-4473-9661-97ac75a187fb. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 808.385349] env[62692]: DEBUG oslo_concurrency.lockutils [req-8c50b1f7-4efa-42ba-99f9-e3b3b334678a req-63e90339-b36a-4daa-bc49-6198ebcde63a service nova] Acquiring lock "refresh_cache-27583cb4-0478-44d8-8970-478682271513" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.385469] env[62692]: DEBUG oslo_concurrency.lockutils [req-8c50b1f7-4efa-42ba-99f9-e3b3b334678a req-63e90339-b36a-4daa-bc49-6198ebcde63a service nova] Acquired lock "refresh_cache-27583cb4-0478-44d8-8970-478682271513" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.385607] env[62692]: DEBUG nova.network.neutron [req-8c50b1f7-4efa-42ba-99f9-e3b3b334678a req-63e90339-b36a-4daa-bc49-6198ebcde63a service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] Refreshing network info cache for port d4991635-5c62-4473-9661-97ac75a187fb {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 808.404699] env[62692]: DEBUG nova.compute.manager [req-5f218c8b-38ef-466a-ba5b-cca521fabe27 req-13345a5c-d9f4-477c-b163-c5bc553fa33a service nova] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Received event network-vif-deleted-469cb44a-0d70-4493-a248-8c3336ebc7b6 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 808.404934] env[62692]: INFO nova.compute.manager [req-5f218c8b-38ef-466a-ba5b-cca521fabe27 req-13345a5c-d9f4-477c-b163-c5bc553fa33a service nova] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Neutron deleted interface 469cb44a-0d70-4493-a248-8c3336ebc7b6; detaching it from the instance and deleting it from the info cache [ 808.405131] env[62692]: DEBUG nova.network.neutron [req-5f218c8b-38ef-466a-ba5b-cca521fabe27 req-13345a5c-d9f4-477c-b163-c5bc553fa33a service nova] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.439586] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141002, 'name': ReconfigVM_Task, 'duration_secs': 0.606537} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.439936] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Reconfigured VM instance instance-0000002a to attach disk [datastore1] ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64/ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 808.440641] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ccefda82-28f0-4d8c-9576-e0bd8b1cc484 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.450464] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 808.450464] env[62692]: value = "task-1141007" [ 808.450464] env[62692]: _type = "Task" [ 808.450464] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.463943] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141007, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.582719] env[62692]: DEBUG oslo_vmware.api [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Task: {'id': task-1141003, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.642058] env[62692]: DEBUG oslo_concurrency.lockutils [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.618s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.642848] env[62692]: DEBUG nova.compute.manager [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 808.646813] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.319s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.647084] env[62692]: DEBUG nova.objects.instance [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Lazy-loading 'resources' on Instance uuid c1818ed2-9e27-4738-bc17-98832318cf61 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 808.691906] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141006, 'name': CreateVM_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.810060] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.844902] env[62692]: DEBUG nova.network.neutron [-] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.857897] env[62692]: DEBUG oslo_concurrency.lockutils [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.909948] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ad60da0f-75b8-4e12-9124-3cb23263d7ee {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.920540] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aee6fa3-8518-459a-a79c-4409f0f50fdd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.957510] env[62692]: DEBUG nova.compute.manager [req-5f218c8b-38ef-466a-ba5b-cca521fabe27 req-13345a5c-d9f4-477c-b163-c5bc553fa33a service nova] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Detach interface failed, port_id=469cb44a-0d70-4493-a248-8c3336ebc7b6, reason: Instance 105e6fd5-4eff-4a0c-9a4b-e6deade781d5 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 808.970255] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141007, 'name': Rename_Task, 'duration_secs': 0.22451} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.970255] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 808.970255] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f3799c10-4e19-46f3-83e0-83ce4d2b1f3c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.976054] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 808.976054] env[62692]: value = "task-1141008" [ 808.976054] env[62692]: _type = "Task" [ 808.976054] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.984444] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141008, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.079025] env[62692]: DEBUG oslo_vmware.api [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Task: {'id': task-1141003, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.777062} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.079370] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 6097d6b2-52d5-4765-94c0-d5f3609464d8/6097d6b2-52d5-4765-94c0-d5f3609464d8.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 809.079651] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 809.079988] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5b64bb24-bafe-4304-bd67-929c557ac135 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.089472] env[62692]: DEBUG oslo_vmware.api [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Waiting for the task: (returnval){ [ 809.089472] env[62692]: value = "task-1141009" [ 809.089472] env[62692]: _type = "Task" [ 809.089472] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.103170] env[62692]: DEBUG oslo_vmware.api [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Task: {'id': task-1141009, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.150716] env[62692]: DEBUG nova.compute.utils [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 809.156272] env[62692]: DEBUG nova.compute.manager [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Not allocating networking since 'none' was specified. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 809.197651] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141006, 'name': CreateVM_Task, 'duration_secs': 0.822416} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.200254] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27583cb4-0478-44d8-8970-478682271513] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 809.201179] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.201365] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.201654] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 809.201912] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7aeb16ba-121f-41d1-a10a-358e8a529264 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.207588] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 809.207588] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]521deb75-b511-f33b-8175-b753c93ea725" [ 809.207588] env[62692]: _type = "Task" [ 809.207588] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.211626] env[62692]: DEBUG nova.network.neutron [req-8c50b1f7-4efa-42ba-99f9-e3b3b334678a req-63e90339-b36a-4daa-bc49-6198ebcde63a service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] Updated VIF entry in instance network info cache for port d4991635-5c62-4473-9661-97ac75a187fb. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 809.211958] env[62692]: DEBUG nova.network.neutron [req-8c50b1f7-4efa-42ba-99f9-e3b3b334678a req-63e90339-b36a-4daa-bc49-6198ebcde63a service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] Updating instance_info_cache with network_info: [{"id": "d4991635-5c62-4473-9661-97ac75a187fb", "address": "fa:16:3e:d4:98:43", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4991635-5c", "ovs_interfaceid": "d4991635-5c62-4473-9661-97ac75a187fb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.226477] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]521deb75-b511-f33b-8175-b753c93ea725, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.350835] env[62692]: INFO nova.compute.manager [-] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Took 1.30 seconds to deallocate network for instance. [ 809.373365] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "fa7bb883-0065-4ecc-9eb5-8c5443344a2a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.373675] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "fa7bb883-0065-4ecc-9eb5-8c5443344a2a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.373891] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "fa7bb883-0065-4ecc-9eb5-8c5443344a2a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.374122] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "fa7bb883-0065-4ecc-9eb5-8c5443344a2a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.374306] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "fa7bb883-0065-4ecc-9eb5-8c5443344a2a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.380742] env[62692]: INFO nova.compute.manager [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Terminating instance [ 809.383015] env[62692]: DEBUG nova.compute.manager [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 809.384922] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 809.384922] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6e94337-6d65-4bac-87aa-399f38975b0b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.393254] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 809.400180] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c8ee712c-1f95-4521-aab4-bf0affae5e89 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.406964] env[62692]: DEBUG oslo_vmware.api [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 809.406964] env[62692]: value = "task-1141010" [ 809.406964] env[62692]: _type = "Task" [ 809.406964] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.420725] env[62692]: DEBUG oslo_vmware.api [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141010, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.459450] env[62692]: DEBUG nova.network.neutron [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Successfully updated port: b0db2db5-00e9-4a3a-a41a-7ed7f4985073 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 809.491977] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141008, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.607204] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4e7dfb8-816a-4999-a0e1-7f792ca672dd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.614316] env[62692]: DEBUG oslo_vmware.api [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Task: {'id': task-1141009, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083043} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.614605] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 809.616092] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a29f3aec-8954-480c-9d55-83a80786bb5a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.622257] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a52313-ca72-42e2-ab18-f7778ff97acc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.644793] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Reconfiguring VM instance instance-0000002b to attach disk [datastore1] 6097d6b2-52d5-4765-94c0-d5f3609464d8/6097d6b2-52d5-4765-94c0-d5f3609464d8.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 809.645737] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0a7ca2f4-12f0-4118-9b0d-f0f2ee1f4929 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.689751] env[62692]: DEBUG nova.compute.manager [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 809.696020] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e27ac05e-979b-4f49-83dc-3d27561b3672 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.699926] env[62692]: DEBUG oslo_vmware.api [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Waiting for the task: (returnval){ [ 809.699926] env[62692]: value = "task-1141011" [ 809.699926] env[62692]: _type = "Task" [ 809.699926] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.706577] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8dfc9d7-801d-4f8c-9837-041fb103f03d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.719672] env[62692]: DEBUG oslo_vmware.api [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Task: {'id': task-1141011, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.733018] env[62692]: DEBUG oslo_concurrency.lockutils [req-8c50b1f7-4efa-42ba-99f9-e3b3b334678a req-63e90339-b36a-4daa-bc49-6198ebcde63a service nova] Releasing lock "refresh_cache-27583cb4-0478-44d8-8970-478682271513" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.733018] env[62692]: DEBUG nova.compute.provider_tree [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 809.743712] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]521deb75-b511-f33b-8175-b753c93ea725, 'name': SearchDatastore_Task, 'duration_secs': 0.035378} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.743937] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.744045] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 809.744298] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.744442] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.744668] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 809.745719] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-52475faf-ba44-4671-a14b-7af782a115b2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.760193] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 809.760193] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 809.760193] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fcdf8f2d-902a-484b-8864-b63b454f12d1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.767550] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 809.767550] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5289f20a-3554-4919-8b85-75b757cc2c9c" [ 809.767550] env[62692]: _type = "Task" [ 809.767550] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.777656] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5289f20a-3554-4919-8b85-75b757cc2c9c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.857225] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.916875] env[62692]: DEBUG oslo_vmware.api [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141010, 'name': PowerOffVM_Task, 'duration_secs': 0.37345} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.917215] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 809.917386] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 809.917733] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8019a752-bc45-45a9-91d5-7934d613e20d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.984607] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 809.984942] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 809.985174] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Deleting the datastore file [datastore1] fa7bb883-0065-4ecc-9eb5-8c5443344a2a {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 809.986076] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7b8a7e61-c93e-4246-b4e1-149af9b403e1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.991275] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141008, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.997267] env[62692]: DEBUG oslo_vmware.api [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 809.997267] env[62692]: value = "task-1141013" [ 809.997267] env[62692]: _type = "Task" [ 809.997267] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.006496] env[62692]: DEBUG oslo_vmware.api [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141013, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.037289] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "433f740c-bde4-4d33-baca-5a7ac8179da1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.037520] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "433f740c-bde4-4d33-baca-5a7ac8179da1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.213185] env[62692]: DEBUG oslo_vmware.api [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Task: {'id': task-1141011, 'name': ReconfigVM_Task, 'duration_secs': 0.459077} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.213708] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Reconfigured VM instance instance-0000002b to attach disk [datastore1] 6097d6b2-52d5-4765-94c0-d5f3609464d8/6097d6b2-52d5-4765-94c0-d5f3609464d8.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 810.214736] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ef7d2921-8048-4ee5-a2bf-bf0838d468b2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.222685] env[62692]: DEBUG oslo_vmware.api [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Waiting for the task: (returnval){ [ 810.222685] env[62692]: value = "task-1141014" [ 810.222685] env[62692]: _type = "Task" [ 810.222685] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.232476] env[62692]: DEBUG oslo_vmware.api [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Task: {'id': task-1141014, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.245472] env[62692]: DEBUG nova.scheduler.client.report [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 810.279904] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5289f20a-3554-4919-8b85-75b757cc2c9c, 'name': SearchDatastore_Task, 'duration_secs': 0.015211} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.281209] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-151efb76-f994-4d0c-b232-7c48631e5d70 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.294042] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 810.294042] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52ab4534-7049-39d6-9a6e-e3c2d03fa42c" [ 810.294042] env[62692]: _type = "Task" [ 810.294042] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.304316] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52ab4534-7049-39d6-9a6e-e3c2d03fa42c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.413818] env[62692]: DEBUG nova.compute.manager [req-f053d3fe-00f0-444d-9547-e08f32a77aae req-29f333d5-a7ea-404f-ba18-31eb5e9972e9 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Received event network-vif-plugged-b0db2db5-00e9-4a3a-a41a-7ed7f4985073 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 810.414190] env[62692]: DEBUG oslo_concurrency.lockutils [req-f053d3fe-00f0-444d-9547-e08f32a77aae req-29f333d5-a7ea-404f-ba18-31eb5e9972e9 service nova] Acquiring lock "aa0666ce-4d4f-4333-8acb-ea0cfb438cd9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.414747] env[62692]: DEBUG oslo_concurrency.lockutils [req-f053d3fe-00f0-444d-9547-e08f32a77aae req-29f333d5-a7ea-404f-ba18-31eb5e9972e9 service nova] Lock "aa0666ce-4d4f-4333-8acb-ea0cfb438cd9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.414747] env[62692]: DEBUG oslo_concurrency.lockutils [req-f053d3fe-00f0-444d-9547-e08f32a77aae req-29f333d5-a7ea-404f-ba18-31eb5e9972e9 service nova] Lock "aa0666ce-4d4f-4333-8acb-ea0cfb438cd9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.414747] env[62692]: DEBUG nova.compute.manager [req-f053d3fe-00f0-444d-9547-e08f32a77aae req-29f333d5-a7ea-404f-ba18-31eb5e9972e9 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] No waiting events found dispatching network-vif-plugged-b0db2db5-00e9-4a3a-a41a-7ed7f4985073 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 810.414747] env[62692]: WARNING nova.compute.manager [req-f053d3fe-00f0-444d-9547-e08f32a77aae req-29f333d5-a7ea-404f-ba18-31eb5e9972e9 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Received unexpected event network-vif-plugged-b0db2db5-00e9-4a3a-a41a-7ed7f4985073 for instance with vm_state building and task_state spawning. [ 810.415018] env[62692]: DEBUG nova.compute.manager [req-f053d3fe-00f0-444d-9547-e08f32a77aae req-29f333d5-a7ea-404f-ba18-31eb5e9972e9 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Received event network-changed-b0db2db5-00e9-4a3a-a41a-7ed7f4985073 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 810.415082] env[62692]: DEBUG nova.compute.manager [req-f053d3fe-00f0-444d-9547-e08f32a77aae req-29f333d5-a7ea-404f-ba18-31eb5e9972e9 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Refreshing instance network info cache due to event network-changed-b0db2db5-00e9-4a3a-a41a-7ed7f4985073. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 810.415249] env[62692]: DEBUG oslo_concurrency.lockutils [req-f053d3fe-00f0-444d-9547-e08f32a77aae req-29f333d5-a7ea-404f-ba18-31eb5e9972e9 service nova] Acquiring lock "refresh_cache-aa0666ce-4d4f-4333-8acb-ea0cfb438cd9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.415380] env[62692]: DEBUG oslo_concurrency.lockutils [req-f053d3fe-00f0-444d-9547-e08f32a77aae req-29f333d5-a7ea-404f-ba18-31eb5e9972e9 service nova] Acquired lock "refresh_cache-aa0666ce-4d4f-4333-8acb-ea0cfb438cd9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.415534] env[62692]: DEBUG nova.network.neutron [req-f053d3fe-00f0-444d-9547-e08f32a77aae req-29f333d5-a7ea-404f-ba18-31eb5e9972e9 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Refreshing network info cache for port b0db2db5-00e9-4a3a-a41a-7ed7f4985073 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 810.487059] env[62692]: DEBUG oslo_vmware.api [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141008, 'name': PowerOnVM_Task, 'duration_secs': 1.338961} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.487377] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 810.487580] env[62692]: INFO nova.compute.manager [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Took 10.23 seconds to spawn the instance on the hypervisor. [ 810.487762] env[62692]: DEBUG nova.compute.manager [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 810.488557] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9307a2f-766f-478c-8eaf-df81de516372 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.509173] env[62692]: DEBUG oslo_vmware.api [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141013, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.280365} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.509429] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 810.509617] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 810.509835] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 810.510027] env[62692]: INFO nova.compute.manager [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Took 1.13 seconds to destroy the instance on the hypervisor. [ 810.510271] env[62692]: DEBUG oslo.service.loopingcall [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 810.510471] env[62692]: DEBUG nova.compute.manager [-] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 810.510566] env[62692]: DEBUG nova.network.neutron [-] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 810.708105] env[62692]: DEBUG nova.compute.manager [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 810.736867] env[62692]: DEBUG oslo_vmware.api [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Task: {'id': task-1141014, 'name': Rename_Task, 'duration_secs': 0.299426} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.739320] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 810.739508] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8c392c03-66b8-4a1c-be00-a4940a7a75b5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.747468] env[62692]: DEBUG nova.virt.hardware [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 810.747748] env[62692]: DEBUG nova.virt.hardware [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 810.748035] env[62692]: DEBUG nova.virt.hardware [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 810.748035] env[62692]: DEBUG nova.virt.hardware [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 810.748249] env[62692]: DEBUG nova.virt.hardware [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 810.748343] env[62692]: DEBUG nova.virt.hardware [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 810.748569] env[62692]: DEBUG nova.virt.hardware [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 810.748745] env[62692]: DEBUG nova.virt.hardware [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 810.748933] env[62692]: DEBUG nova.virt.hardware [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 810.749110] env[62692]: DEBUG nova.virt.hardware [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 810.749285] env[62692]: DEBUG nova.virt.hardware [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 810.750148] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e69622dd-e741-4af7-ac5c-f19b6a5873fe {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.754768] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.108s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.756783] env[62692]: DEBUG oslo_vmware.api [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Waiting for the task: (returnval){ [ 810.756783] env[62692]: value = "task-1141015" [ 810.756783] env[62692]: _type = "Task" [ 810.756783] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.758771] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.804s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.762236] env[62692]: INFO nova.compute.claims [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 810.772477] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf99c274-1420-4282-a071-66fa1cb94a19 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.792308] env[62692]: DEBUG oslo_vmware.api [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Task: {'id': task-1141015, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.799243] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Instance VIF info [] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 810.811538] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Creating folder: Project (e06a5ebaf3b04303aa48bb198a145dfd). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 810.811538] env[62692]: INFO nova.scheduler.client.report [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Deleted allocations for instance c1818ed2-9e27-4738-bc17-98832318cf61 [ 810.821105] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-810ca177-fc0f-4e8c-9363-8038aa49fae9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.830450] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52ab4534-7049-39d6-9a6e-e3c2d03fa42c, 'name': SearchDatastore_Task, 'duration_secs': 0.016039} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.830664] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.830944] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 27583cb4-0478-44d8-8970-478682271513/27583cb4-0478-44d8-8970-478682271513.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 810.831265] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c476d7df-b91c-4995-9656-f9ec55637c98 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.835443] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Created folder: Project (e06a5ebaf3b04303aa48bb198a145dfd) in parent group-v248868. [ 810.835622] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Creating folder: Instances. Parent ref: group-v248913. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 810.837638] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9da86d2f-f945-4041-a024-3a2cb18a9833 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.843328] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 810.843328] env[62692]: value = "task-1141017" [ 810.843328] env[62692]: _type = "Task" [ 810.843328] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.854803] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141017, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.856484] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Created folder: Instances in parent group-v248913. [ 810.856926] env[62692]: DEBUG oslo.service.loopingcall [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 810.857167] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 810.857378] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-795d7127-9da4-42ba-9c23-4e12364f1b49 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.877629] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 810.877629] env[62692]: value = "task-1141019" [ 810.877629] env[62692]: _type = "Task" [ 810.877629] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.888018] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141019, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.895363] env[62692]: DEBUG nova.compute.manager [req-20c84d2d-0f7e-4350-af86-344214ae90ae req-f1354e5c-5091-460b-84b3-e03a7da0988d service nova] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Received event network-vif-deleted-f19989c4-0ce9-47ae-989b-0faab24bc9b1 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 810.895547] env[62692]: INFO nova.compute.manager [req-20c84d2d-0f7e-4350-af86-344214ae90ae req-f1354e5c-5091-460b-84b3-e03a7da0988d service nova] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Neutron deleted interface f19989c4-0ce9-47ae-989b-0faab24bc9b1; detaching it from the instance and deleting it from the info cache [ 810.895802] env[62692]: DEBUG nova.network.neutron [req-20c84d2d-0f7e-4350-af86-344214ae90ae req-f1354e5c-5091-460b-84b3-e03a7da0988d service nova] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.977665] env[62692]: DEBUG nova.network.neutron [req-f053d3fe-00f0-444d-9547-e08f32a77aae req-29f333d5-a7ea-404f-ba18-31eb5e9972e9 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 811.012616] env[62692]: INFO nova.compute.manager [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Took 28.68 seconds to build instance. [ 811.116158] env[62692]: DEBUG nova.network.neutron [req-f053d3fe-00f0-444d-9547-e08f32a77aae req-29f333d5-a7ea-404f-ba18-31eb5e9972e9 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.278217] env[62692]: DEBUG oslo_vmware.api [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Task: {'id': task-1141015, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.327785] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e71e7ea1-2419-4640-aabd-c419e18d9729 tempest-ServersAaction247Test-1455060593 tempest-ServersAaction247Test-1455060593-project-member] Lock "c1818ed2-9e27-4738-bc17-98832318cf61" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.103s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.341266] env[62692]: DEBUG nova.network.neutron [-] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.356197] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141017, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.395940] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141019, 'name': CreateVM_Task, 'duration_secs': 0.327772} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.396215] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 811.396980] env[62692]: DEBUG oslo_concurrency.lockutils [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.397143] env[62692]: DEBUG oslo_concurrency.lockutils [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.397688] env[62692]: DEBUG oslo_concurrency.lockutils [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 811.398413] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc13b0bc-034e-47bf-af89-9d5cf477e1db {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.401646] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-242018c4-fa96-4961-a7fb-2ef06de8b353 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.415211] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc213aa1-5bb9-43f7-96e1-3cf2e0dee198 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.426800] env[62692]: DEBUG oslo_vmware.api [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Waiting for the task: (returnval){ [ 811.426800] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52667336-1bcf-2f47-038b-73141b5ba5b4" [ 811.426800] env[62692]: _type = "Task" [ 811.426800] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.440330] env[62692]: INFO nova.compute.manager [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Rescuing [ 811.440778] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "refresh_cache-ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.441043] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquired lock "refresh_cache-ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.441313] env[62692]: DEBUG nova.network.neutron [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 811.443072] env[62692]: DEBUG oslo_vmware.api [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52667336-1bcf-2f47-038b-73141b5ba5b4, 'name': SearchDatastore_Task, 'duration_secs': 0.023942} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.443761] env[62692]: DEBUG oslo_concurrency.lockutils [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.444397] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 811.444756] env[62692]: DEBUG oslo_concurrency.lockutils [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.445049] env[62692]: DEBUG oslo_concurrency.lockutils [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.445355] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 811.445828] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f38699d5-5a68-492c-825a-cf2ea9ef495d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.458804] env[62692]: DEBUG nova.compute.manager [req-20c84d2d-0f7e-4350-af86-344214ae90ae req-f1354e5c-5091-460b-84b3-e03a7da0988d service nova] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Detach interface failed, port_id=f19989c4-0ce9-47ae-989b-0faab24bc9b1, reason: Instance fa7bb883-0065-4ecc-9eb5-8c5443344a2a could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 811.469661] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 811.469880] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 811.470674] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21590d91-a34a-4b17-a817-a8a29e1cd9a2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.477313] env[62692]: DEBUG oslo_vmware.api [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Waiting for the task: (returnval){ [ 811.477313] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52846a23-6ca2-3bbf-1f45-63ce82ab7e23" [ 811.477313] env[62692]: _type = "Task" [ 811.477313] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.488029] env[62692]: DEBUG oslo_vmware.api [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52846a23-6ca2-3bbf-1f45-63ce82ab7e23, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.515624] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f2bfc11-cb8b-4630-ae5a-adb6f0be9bbc tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 149.537s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.621256] env[62692]: DEBUG oslo_concurrency.lockutils [req-f053d3fe-00f0-444d-9547-e08f32a77aae req-29f333d5-a7ea-404f-ba18-31eb5e9972e9 service nova] Releasing lock "refresh_cache-aa0666ce-4d4f-4333-8acb-ea0cfb438cd9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.723733] env[62692]: DEBUG nova.network.neutron [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Successfully updated port: 13349e97-1acd-48d2-a24b-44160a60773e {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 811.773788] env[62692]: DEBUG oslo_vmware.api [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Task: {'id': task-1141015, 'name': PowerOnVM_Task, 'duration_secs': 0.665994} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.773903] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 811.774108] env[62692]: INFO nova.compute.manager [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Took 8.92 seconds to spawn the instance on the hypervisor. [ 811.774312] env[62692]: DEBUG nova.compute.manager [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 811.775158] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d83a693f-6516-4d08-80f5-803ea8d6ad2f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.845131] env[62692]: INFO nova.compute.manager [-] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Took 1.33 seconds to deallocate network for instance. [ 811.859984] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141017, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.628151} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.862727] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 27583cb4-0478-44d8-8970-478682271513/27583cb4-0478-44d8-8970-478682271513.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 811.863505] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 811.863901] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8503289b-e93e-4e3d-bc51-a24b828a6ba7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.872914] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 811.872914] env[62692]: value = "task-1141020" [ 811.872914] env[62692]: _type = "Task" [ 811.872914] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.892123] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141020, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.991527] env[62692]: DEBUG oslo_vmware.api [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52846a23-6ca2-3bbf-1f45-63ce82ab7e23, 'name': SearchDatastore_Task, 'duration_secs': 0.016552} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.993137] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ba0f538-801b-4072-9b77-cdfe5a1455f1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.001283] env[62692]: DEBUG oslo_vmware.api [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Waiting for the task: (returnval){ [ 812.001283] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52aa574a-3622-0db9-c313-87864540c6d9" [ 812.001283] env[62692]: _type = "Task" [ 812.001283] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.013642] env[62692]: DEBUG oslo_vmware.api [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52aa574a-3622-0db9-c313-87864540c6d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.018409] env[62692]: DEBUG nova.compute.manager [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 812.173456] env[62692]: DEBUG nova.network.neutron [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Updating instance_info_cache with network_info: [{"id": "920fbe7b-7ba1-4d33-bf4f-d96609ed6640", "address": "fa:16:3e:5e:35:e2", "network": {"id": "17e72c3b-603f-4305-96ac-89a9b4c9c7e9", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-123501993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7db5465cb5db465b92dfa570572b6e80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap920fbe7b-7b", "ovs_interfaceid": "920fbe7b-7ba1-4d33-bf4f-d96609ed6640", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.218952] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b80db651-e8a3-4291-bd34-f9dcc3694d51 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.229414] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-565184b3-52db-47ee-b83d-13ccbcd04816 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.265138] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edfd37c2-d10d-4f25-9a5f-f848445dda11 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.273433] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a127f4-bbfa-401e-a776-bf6e99f5ce17 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.288622] env[62692]: DEBUG nova.compute.provider_tree [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 812.298892] env[62692]: INFO nova.compute.manager [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Took 28.33 seconds to build instance. [ 812.354405] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.383929] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141020, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076171} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.384222] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 812.385080] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83612821-04dc-45e2-ab16-56dbc42ead55 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.408173] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Reconfiguring VM instance instance-0000002c to attach disk [datastore1] 27583cb4-0478-44d8-8970-478682271513/27583cb4-0478-44d8-8970-478682271513.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 812.408456] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f795bfc1-807e-4e7e-ae9d-c81c9fbd9f95 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.430389] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 812.430389] env[62692]: value = "task-1141021" [ 812.430389] env[62692]: _type = "Task" [ 812.430389] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.444384] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141021, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.512267] env[62692]: DEBUG oslo_vmware.api [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52aa574a-3622-0db9-c313-87864540c6d9, 'name': SearchDatastore_Task, 'duration_secs': 0.047746} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.512564] env[62692]: DEBUG oslo_concurrency.lockutils [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.512934] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 1848c251-0a0c-4c36-8bd0-ff37befde2a0/1848c251-0a0c-4c36-8bd0-ff37befde2a0.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 812.513090] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-766ad597-a938-4ca6-8c62-a3784cf71f1d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.520560] env[62692]: DEBUG oslo_vmware.api [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Waiting for the task: (returnval){ [ 812.520560] env[62692]: value = "task-1141022" [ 812.520560] env[62692]: _type = "Task" [ 812.520560] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.531874] env[62692]: DEBUG oslo_vmware.api [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141022, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.545060] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.654547] env[62692]: DEBUG nova.compute.manager [req-f27fcaf9-4565-4388-98c5-198c67dcbf94 req-ac9fcfdc-6f9f-41dd-819b-2d4f8da3cd6f service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Received event network-vif-plugged-13349e97-1acd-48d2-a24b-44160a60773e {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 812.654831] env[62692]: DEBUG oslo_concurrency.lockutils [req-f27fcaf9-4565-4388-98c5-198c67dcbf94 req-ac9fcfdc-6f9f-41dd-819b-2d4f8da3cd6f service nova] Acquiring lock "aa0666ce-4d4f-4333-8acb-ea0cfb438cd9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.654968] env[62692]: DEBUG oslo_concurrency.lockutils [req-f27fcaf9-4565-4388-98c5-198c67dcbf94 req-ac9fcfdc-6f9f-41dd-819b-2d4f8da3cd6f service nova] Lock "aa0666ce-4d4f-4333-8acb-ea0cfb438cd9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.655276] env[62692]: DEBUG oslo_concurrency.lockutils [req-f27fcaf9-4565-4388-98c5-198c67dcbf94 req-ac9fcfdc-6f9f-41dd-819b-2d4f8da3cd6f service nova] Lock "aa0666ce-4d4f-4333-8acb-ea0cfb438cd9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.655389] env[62692]: DEBUG nova.compute.manager [req-f27fcaf9-4565-4388-98c5-198c67dcbf94 req-ac9fcfdc-6f9f-41dd-819b-2d4f8da3cd6f service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] No waiting events found dispatching network-vif-plugged-13349e97-1acd-48d2-a24b-44160a60773e {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 812.655525] env[62692]: WARNING nova.compute.manager [req-f27fcaf9-4565-4388-98c5-198c67dcbf94 req-ac9fcfdc-6f9f-41dd-819b-2d4f8da3cd6f service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Received unexpected event network-vif-plugged-13349e97-1acd-48d2-a24b-44160a60773e for instance with vm_state building and task_state spawning. [ 812.655635] env[62692]: DEBUG nova.compute.manager [req-f27fcaf9-4565-4388-98c5-198c67dcbf94 req-ac9fcfdc-6f9f-41dd-819b-2d4f8da3cd6f service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Received event network-changed-13349e97-1acd-48d2-a24b-44160a60773e {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 812.655819] env[62692]: DEBUG nova.compute.manager [req-f27fcaf9-4565-4388-98c5-198c67dcbf94 req-ac9fcfdc-6f9f-41dd-819b-2d4f8da3cd6f service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Refreshing instance network info cache due to event network-changed-13349e97-1acd-48d2-a24b-44160a60773e. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 812.655991] env[62692]: DEBUG oslo_concurrency.lockutils [req-f27fcaf9-4565-4388-98c5-198c67dcbf94 req-ac9fcfdc-6f9f-41dd-819b-2d4f8da3cd6f service nova] Acquiring lock "refresh_cache-aa0666ce-4d4f-4333-8acb-ea0cfb438cd9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.656140] env[62692]: DEBUG oslo_concurrency.lockutils [req-f27fcaf9-4565-4388-98c5-198c67dcbf94 req-ac9fcfdc-6f9f-41dd-819b-2d4f8da3cd6f service nova] Acquired lock "refresh_cache-aa0666ce-4d4f-4333-8acb-ea0cfb438cd9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.656305] env[62692]: DEBUG nova.network.neutron [req-f27fcaf9-4565-4388-98c5-198c67dcbf94 req-ac9fcfdc-6f9f-41dd-819b-2d4f8da3cd6f service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Refreshing network info cache for port 13349e97-1acd-48d2-a24b-44160a60773e {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 812.677903] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Releasing lock "refresh_cache-ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.795179] env[62692]: DEBUG nova.scheduler.client.report [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 812.798821] env[62692]: DEBUG oslo_concurrency.lockutils [None req-33f47aea-3fa1-40a9-b116-d7cd35cd576a tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Lock "6097d6b2-52d5-4765-94c0-d5f3609464d8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.378s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.947955] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141021, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.035979] env[62692]: DEBUG oslo_vmware.api [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141022, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.204021] env[62692]: DEBUG nova.network.neutron [req-f27fcaf9-4565-4388-98c5-198c67dcbf94 req-ac9fcfdc-6f9f-41dd-819b-2d4f8da3cd6f service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 813.213475] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 813.213775] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cc3d2a28-f852-418e-8798-72d86bb5b184 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.222886] env[62692]: DEBUG oslo_vmware.api [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 813.222886] env[62692]: value = "task-1141023" [ 813.222886] env[62692]: _type = "Task" [ 813.222886] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.232421] env[62692]: DEBUG oslo_vmware.api [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141023, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.300946] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.543s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.301440] env[62692]: DEBUG nova.compute.manager [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 813.304494] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.170s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.306519] env[62692]: INFO nova.compute.claims [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 813.309379] env[62692]: DEBUG nova.compute.manager [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 813.348531] env[62692]: DEBUG nova.network.neutron [req-f27fcaf9-4565-4388-98c5-198c67dcbf94 req-ac9fcfdc-6f9f-41dd-819b-2d4f8da3cd6f service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.445579] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141021, 'name': ReconfigVM_Task, 'duration_secs': 0.597071} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.446080] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Reconfigured VM instance instance-0000002c to attach disk [datastore1] 27583cb4-0478-44d8-8970-478682271513/27583cb4-0478-44d8-8970-478682271513.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 813.446929] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e899cd4e-00a1-4acb-be30-8d229030ad0c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.455049] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 813.455049] env[62692]: value = "task-1141024" [ 813.455049] env[62692]: _type = "Task" [ 813.455049] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.464725] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141024, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.532370] env[62692]: DEBUG oslo_vmware.api [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141022, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.783526} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.532642] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 1848c251-0a0c-4c36-8bd0-ff37befde2a0/1848c251-0a0c-4c36-8bd0-ff37befde2a0.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 813.532857] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 813.533120] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1664a9e7-d680-43e0-a758-b1a5107fb5da {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.542590] env[62692]: DEBUG oslo_vmware.api [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Waiting for the task: (returnval){ [ 813.542590] env[62692]: value = "task-1141025" [ 813.542590] env[62692]: _type = "Task" [ 813.542590] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.552637] env[62692]: DEBUG oslo_vmware.api [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141025, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.645126] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Acquiring lock "6097d6b2-52d5-4765-94c0-d5f3609464d8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.645531] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Lock "6097d6b2-52d5-4765-94c0-d5f3609464d8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.645782] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Acquiring lock "6097d6b2-52d5-4765-94c0-d5f3609464d8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.645985] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Lock "6097d6b2-52d5-4765-94c0-d5f3609464d8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.646175] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Lock "6097d6b2-52d5-4765-94c0-d5f3609464d8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.648672] env[62692]: INFO nova.compute.manager [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Terminating instance [ 813.650770] env[62692]: DEBUG nova.compute.manager [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 813.651032] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 813.651912] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fbe3b4a-e01b-4292-9a25-48a3308ce709 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.660783] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 813.660995] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-313e71b6-2de1-4d2f-b5c0-7c8d81f2f27b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.668841] env[62692]: DEBUG oslo_vmware.api [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Waiting for the task: (returnval){ [ 813.668841] env[62692]: value = "task-1141026" [ 813.668841] env[62692]: _type = "Task" [ 813.668841] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.681376] env[62692]: DEBUG oslo_vmware.api [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Task: {'id': task-1141026, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.733708] env[62692]: DEBUG oslo_vmware.api [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141023, 'name': PowerOffVM_Task, 'duration_secs': 0.412611} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.734053] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 813.734788] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2b409b6-7362-4af7-acca-f962c599b705 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.756168] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3790b6f-8022-4ba0-8ca0-916f65582dee {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.794821] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 813.795152] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6de13b3b-17f8-4532-a026-8d4d362896b6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.806794] env[62692]: DEBUG oslo_vmware.api [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 813.806794] env[62692]: value = "task-1141027" [ 813.806794] env[62692]: _type = "Task" [ 813.806794] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.817967] env[62692]: DEBUG nova.compute.utils [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 813.825858] env[62692]: DEBUG nova.compute.manager [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 813.826318] env[62692]: DEBUG nova.network.neutron [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 813.830073] env[62692]: DEBUG oslo_vmware.api [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141027, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.854415] env[62692]: DEBUG oslo_concurrency.lockutils [req-f27fcaf9-4565-4388-98c5-198c67dcbf94 req-ac9fcfdc-6f9f-41dd-819b-2d4f8da3cd6f service nova] Releasing lock "refresh_cache-aa0666ce-4d4f-4333-8acb-ea0cfb438cd9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.854683] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.903111] env[62692]: DEBUG nova.policy [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c70223af914c4c99b588f84940c8545c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '198a5af1f19c49e8a3b070cfbe9ee01a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 813.968098] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141024, 'name': Rename_Task, 'duration_secs': 0.179128} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.968443] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 813.968729] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ab11a546-81d9-4c49-83db-6cad7e4dc338 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.980020] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 813.980020] env[62692]: value = "task-1141028" [ 813.980020] env[62692]: _type = "Task" [ 813.980020] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.993474] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141028, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.996616] env[62692]: DEBUG nova.network.neutron [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Successfully updated port: 56f47df0-50f1-43e0-a2d5-3a4ad9d19327 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 814.054981] env[62692]: DEBUG oslo_vmware.api [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141025, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.102401} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.054981] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 814.054981] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-765c08d5-84e2-4f50-bec1-22d85d31a880 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.077758] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Reconfiguring VM instance instance-0000002e to attach disk [datastore2] 1848c251-0a0c-4c36-8bd0-ff37befde2a0/1848c251-0a0c-4c36-8bd0-ff37befde2a0.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 814.078542] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4677b3b6-7949-43f5-8338-5ac2f3406496 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.101870] env[62692]: DEBUG oslo_vmware.api [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Waiting for the task: (returnval){ [ 814.101870] env[62692]: value = "task-1141029" [ 814.101870] env[62692]: _type = "Task" [ 814.101870] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.111876] env[62692]: DEBUG oslo_vmware.api [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141029, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.180168] env[62692]: DEBUG oslo_vmware.api [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Task: {'id': task-1141026, 'name': PowerOffVM_Task, 'duration_secs': 0.184559} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.180472] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 814.180640] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 814.180897] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b7cb4c28-fe2f-4f3a-8f94-154912707def {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.264702] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 814.265110] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 814.265346] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Deleting the datastore file [datastore1] 6097d6b2-52d5-4765-94c0-d5f3609464d8 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 814.265658] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-38714b7b-b7be-458f-bd8b-64528dc4033c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.276506] env[62692]: DEBUG oslo_vmware.api [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Waiting for the task: (returnval){ [ 814.276506] env[62692]: value = "task-1141031" [ 814.276506] env[62692]: _type = "Task" [ 814.276506] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.288083] env[62692]: DEBUG oslo_vmware.api [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Task: {'id': task-1141031, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.302550] env[62692]: DEBUG nova.network.neutron [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Successfully created port: 6b631aba-53ff-466f-b003-cc92a2b87144 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 814.320207] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] VM already powered off {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 814.320207] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 814.321046] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.321046] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.321046] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 814.321046] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ce898a72-e9e1-43a5-a7ab-d8b2db6e0486 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.326802] env[62692]: DEBUG nova.compute.manager [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 814.341605] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 814.341788] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 814.342567] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed0cb03e-73c1-4ea9-b1fb-cb55ed5803a4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.358772] env[62692]: DEBUG oslo_vmware.api [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 814.358772] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]522cdcfe-e387-6392-3bcb-6f169269b2c6" [ 814.358772] env[62692]: _type = "Task" [ 814.358772] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.370760] env[62692]: DEBUG oslo_vmware.api [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]522cdcfe-e387-6392-3bcb-6f169269b2c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.493693] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141028, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.501727] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Acquiring lock "refresh_cache-aa0666ce-4d4f-4333-8acb-ea0cfb438cd9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.501801] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Acquired lock "refresh_cache-aa0666ce-4d4f-4333-8acb-ea0cfb438cd9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.501985] env[62692]: DEBUG nova.network.neutron [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 814.620813] env[62692]: DEBUG oslo_vmware.api [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141029, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.786626] env[62692]: DEBUG oslo_vmware.api [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Task: {'id': task-1141031, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.326844} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.791808] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 814.791808] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 814.791808] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 814.791808] env[62692]: INFO nova.compute.manager [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Took 1.14 seconds to destroy the instance on the hypervisor. [ 814.791808] env[62692]: DEBUG oslo.service.loopingcall [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 814.791996] env[62692]: DEBUG nova.compute.manager [-] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 814.791996] env[62692]: DEBUG nova.network.neutron [-] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 814.836132] env[62692]: INFO nova.virt.block_device [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Booting with volume 9886d0c7-5df2-4bce-8853-faa1431b1865 at /dev/sda [ 814.870739] env[62692]: DEBUG oslo_vmware.api [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]522cdcfe-e387-6392-3bcb-6f169269b2c6, 'name': SearchDatastore_Task, 'duration_secs': 0.02296} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.872008] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-65e2af6b-3800-4362-bc61-0b4c04827d2f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.881088] env[62692]: DEBUG oslo_vmware.api [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 814.881088] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5248dc56-73e8-b961-c809-f0fd1b6ef924" [ 814.881088] env[62692]: _type = "Task" [ 814.881088] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.885470] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2f0afd2-456c-4347-b2fe-9c5eec5a0ce2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.895511] env[62692]: DEBUG oslo_vmware.api [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5248dc56-73e8-b961-c809-f0fd1b6ef924, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.898274] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aae8a522-7f93-4926-98b8-5c837aee1067 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.901849] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9f25d96b-b8d2-421b-b161-b9f688fa21d4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.910556] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db37960e-9e7e-4df7-bb98-3cc07ad49b2c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.949811] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d51b6cf-7e82-4996-879c-3992336756ef {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.957770] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ff75710-d342-4f9c-a91f-d804c140b9d5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.970185] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0c838173-e40d-4c69-8f61-1ec021785078 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.973929] env[62692]: DEBUG nova.compute.manager [req-f4734622-b124-422c-9685-cbec71f22d0d req-c439e92c-bc06-4ddb-a154-9f47d2dbb7f8 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Received event network-vif-plugged-56f47df0-50f1-43e0-a2d5-3a4ad9d19327 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 814.974142] env[62692]: DEBUG oslo_concurrency.lockutils [req-f4734622-b124-422c-9685-cbec71f22d0d req-c439e92c-bc06-4ddb-a154-9f47d2dbb7f8 service nova] Acquiring lock "aa0666ce-4d4f-4333-8acb-ea0cfb438cd9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.974419] env[62692]: DEBUG oslo_concurrency.lockutils [req-f4734622-b124-422c-9685-cbec71f22d0d req-c439e92c-bc06-4ddb-a154-9f47d2dbb7f8 service nova] Lock "aa0666ce-4d4f-4333-8acb-ea0cfb438cd9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.974622] env[62692]: DEBUG oslo_concurrency.lockutils [req-f4734622-b124-422c-9685-cbec71f22d0d req-c439e92c-bc06-4ddb-a154-9f47d2dbb7f8 service nova] Lock "aa0666ce-4d4f-4333-8acb-ea0cfb438cd9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.974806] env[62692]: DEBUG nova.compute.manager [req-f4734622-b124-422c-9685-cbec71f22d0d req-c439e92c-bc06-4ddb-a154-9f47d2dbb7f8 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] No waiting events found dispatching network-vif-plugged-56f47df0-50f1-43e0-a2d5-3a4ad9d19327 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 814.974993] env[62692]: WARNING nova.compute.manager [req-f4734622-b124-422c-9685-cbec71f22d0d req-c439e92c-bc06-4ddb-a154-9f47d2dbb7f8 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Received unexpected event network-vif-plugged-56f47df0-50f1-43e0-a2d5-3a4ad9d19327 for instance with vm_state building and task_state spawning. [ 814.975200] env[62692]: DEBUG nova.compute.manager [req-f4734622-b124-422c-9685-cbec71f22d0d req-c439e92c-bc06-4ddb-a154-9f47d2dbb7f8 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Received event network-changed-56f47df0-50f1-43e0-a2d5-3a4ad9d19327 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 814.975360] env[62692]: DEBUG nova.compute.manager [req-f4734622-b124-422c-9685-cbec71f22d0d req-c439e92c-bc06-4ddb-a154-9f47d2dbb7f8 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Refreshing instance network info cache due to event network-changed-56f47df0-50f1-43e0-a2d5-3a4ad9d19327. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 814.975547] env[62692]: DEBUG oslo_concurrency.lockutils [req-f4734622-b124-422c-9685-cbec71f22d0d req-c439e92c-bc06-4ddb-a154-9f47d2dbb7f8 service nova] Acquiring lock "refresh_cache-aa0666ce-4d4f-4333-8acb-ea0cfb438cd9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.984680] env[62692]: DEBUG nova.compute.provider_tree [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 814.994119] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1f03f12-0af1-4c4b-8d7c-34488756d8bd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.011233] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141028, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.033768] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-179533f4-7547-4f69-b221-68b96277136b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.041740] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a4e3352-16ee-4a2c-88d1-7ecdfe698163 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.057992] env[62692]: DEBUG nova.virt.block_device [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Updating existing volume attachment record: 9a1ab797-e113-46a3-adbf-8248a4b72f20 {{(pid=62692) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 815.067709] env[62692]: DEBUG nova.network.neutron [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 815.114881] env[62692]: DEBUG oslo_vmware.api [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141029, 'name': ReconfigVM_Task, 'duration_secs': 0.578822} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.115202] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Reconfigured VM instance instance-0000002e to attach disk [datastore2] 1848c251-0a0c-4c36-8bd0-ff37befde2a0/1848c251-0a0c-4c36-8bd0-ff37befde2a0.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 815.115869] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dcbe258b-ccb1-4313-9fea-e52a6939d07f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.124140] env[62692]: DEBUG oslo_vmware.api [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Waiting for the task: (returnval){ [ 815.124140] env[62692]: value = "task-1141032" [ 815.124140] env[62692]: _type = "Task" [ 815.124140] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.132950] env[62692]: DEBUG oslo_vmware.api [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141032, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.392051] env[62692]: DEBUG oslo_vmware.api [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5248dc56-73e8-b961-c809-f0fd1b6ef924, 'name': SearchDatastore_Task, 'duration_secs': 0.016104} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.392583] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.392976] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64/81eb64fb-5a0c-47b7-9948-7dcb01b911bd-rescue.vmdk. {{(pid=62692) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 815.393419] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fc8697bc-5488-4fa4-b16d-146bf3b783cd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.401436] env[62692]: DEBUG oslo_vmware.api [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 815.401436] env[62692]: value = "task-1141033" [ 815.401436] env[62692]: _type = "Task" [ 815.401436] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.414088] env[62692]: DEBUG oslo_vmware.api [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141033, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.491692] env[62692]: DEBUG nova.scheduler.client.report [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 815.502044] env[62692]: DEBUG oslo_vmware.api [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141028, 'name': PowerOnVM_Task, 'duration_secs': 1.407593} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.502044] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 815.502044] env[62692]: INFO nova.compute.manager [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Took 10.02 seconds to spawn the instance on the hypervisor. [ 815.502044] env[62692]: DEBUG nova.compute.manager [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 815.502781] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fb27fa0-f93c-411a-b3db-f6bb2c9c43e5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.592462] env[62692]: DEBUG nova.network.neutron [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Updating instance_info_cache with network_info: [{"id": "b0db2db5-00e9-4a3a-a41a-7ed7f4985073", "address": "fa:16:3e:98:ea:2a", "network": {"id": "5f63a59b-8a80-45bb-8b7f-23bfc180fe98", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1827914724", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.125", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fed583a514c4043a4b5772c73e21aff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d12aff80-9d1b-4a67-a470-9c0148b443e3", "external-id": "nsx-vlan-transportzone-784", "segmentation_id": 784, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0db2db5-00", "ovs_interfaceid": "b0db2db5-00e9-4a3a-a41a-7ed7f4985073", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "13349e97-1acd-48d2-a24b-44160a60773e", "address": "fa:16:3e:b7:2e:8e", "network": {"id": "6ad813f0-48dd-45af-89d3-62d3ee8d084f", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-738939750", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.66", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4fed583a514c4043a4b5772c73e21aff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "233536d0-6913-4879-8442-42dcf1d4ecbb", "external-id": "nsx-vlan-transportzone-700", "segmentation_id": 700, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13349e97-1a", "ovs_interfaceid": "13349e97-1acd-48d2-a24b-44160a60773e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "56f47df0-50f1-43e0-a2d5-3a4ad9d19327", "address": "fa:16:3e:35:ae:c9", "network": {"id": "5f63a59b-8a80-45bb-8b7f-23bfc180fe98", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1827914724", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.174", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fed583a514c4043a4b5772c73e21aff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d12aff80-9d1b-4a67-a470-9c0148b443e3", "external-id": "nsx-vlan-transportzone-784", "segmentation_id": 784, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56f47df0-50", "ovs_interfaceid": "56f47df0-50f1-43e0-a2d5-3a4ad9d19327", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.637328] env[62692]: DEBUG oslo_vmware.api [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141032, 'name': Rename_Task, 'duration_secs': 0.186393} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.637811] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 815.638529] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cafc616c-c5f7-4088-a65a-240f3fdf45cb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.647526] env[62692]: DEBUG oslo_vmware.api [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Waiting for the task: (returnval){ [ 815.647526] env[62692]: value = "task-1141034" [ 815.647526] env[62692]: _type = "Task" [ 815.647526] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.656534] env[62692]: DEBUG oslo_vmware.api [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141034, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.719390] env[62692]: DEBUG nova.network.neutron [-] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.917774] env[62692]: DEBUG oslo_vmware.api [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141033, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.969740] env[62692]: DEBUG nova.compute.manager [req-8258ead6-147d-4fa0-9314-b299ae00c242 req-07630150-ce0c-4068-9ff9-f755fe20d961 service nova] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Received event network-vif-plugged-6b631aba-53ff-466f-b003-cc92a2b87144 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 815.969995] env[62692]: DEBUG oslo_concurrency.lockutils [req-8258ead6-147d-4fa0-9314-b299ae00c242 req-07630150-ce0c-4068-9ff9-f755fe20d961 service nova] Acquiring lock "81ce340c-fbef-4932-983c-595843530dbc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.970225] env[62692]: DEBUG oslo_concurrency.lockutils [req-8258ead6-147d-4fa0-9314-b299ae00c242 req-07630150-ce0c-4068-9ff9-f755fe20d961 service nova] Lock "81ce340c-fbef-4932-983c-595843530dbc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.970567] env[62692]: DEBUG oslo_concurrency.lockutils [req-8258ead6-147d-4fa0-9314-b299ae00c242 req-07630150-ce0c-4068-9ff9-f755fe20d961 service nova] Lock "81ce340c-fbef-4932-983c-595843530dbc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.970719] env[62692]: DEBUG nova.compute.manager [req-8258ead6-147d-4fa0-9314-b299ae00c242 req-07630150-ce0c-4068-9ff9-f755fe20d961 service nova] [instance: 81ce340c-fbef-4932-983c-595843530dbc] No waiting events found dispatching network-vif-plugged-6b631aba-53ff-466f-b003-cc92a2b87144 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 815.970885] env[62692]: WARNING nova.compute.manager [req-8258ead6-147d-4fa0-9314-b299ae00c242 req-07630150-ce0c-4068-9ff9-f755fe20d961 service nova] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Received unexpected event network-vif-plugged-6b631aba-53ff-466f-b003-cc92a2b87144 for instance with vm_state building and task_state block_device_mapping. [ 815.997254] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.693s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.998567] env[62692]: DEBUG oslo_concurrency.lockutils [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.575s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.000136] env[62692]: INFO nova.compute.claims [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 816.026874] env[62692]: INFO nova.compute.manager [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Took 22.71 seconds to build instance. [ 816.080571] env[62692]: DEBUG nova.network.neutron [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Successfully updated port: 6b631aba-53ff-466f-b003-cc92a2b87144 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 816.095609] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Releasing lock "refresh_cache-aa0666ce-4d4f-4333-8acb-ea0cfb438cd9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.095609] env[62692]: DEBUG nova.compute.manager [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Instance network_info: |[{"id": "b0db2db5-00e9-4a3a-a41a-7ed7f4985073", "address": "fa:16:3e:98:ea:2a", "network": {"id": "5f63a59b-8a80-45bb-8b7f-23bfc180fe98", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1827914724", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.125", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fed583a514c4043a4b5772c73e21aff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d12aff80-9d1b-4a67-a470-9c0148b443e3", "external-id": "nsx-vlan-transportzone-784", "segmentation_id": 784, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0db2db5-00", "ovs_interfaceid": "b0db2db5-00e9-4a3a-a41a-7ed7f4985073", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "13349e97-1acd-48d2-a24b-44160a60773e", "address": "fa:16:3e:b7:2e:8e", "network": {"id": "6ad813f0-48dd-45af-89d3-62d3ee8d084f", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-738939750", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.66", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4fed583a514c4043a4b5772c73e21aff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "233536d0-6913-4879-8442-42dcf1d4ecbb", "external-id": "nsx-vlan-transportzone-700", "segmentation_id": 700, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13349e97-1a", "ovs_interfaceid": "13349e97-1acd-48d2-a24b-44160a60773e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "56f47df0-50f1-43e0-a2d5-3a4ad9d19327", "address": "fa:16:3e:35:ae:c9", "network": {"id": "5f63a59b-8a80-45bb-8b7f-23bfc180fe98", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1827914724", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.174", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fed583a514c4043a4b5772c73e21aff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d12aff80-9d1b-4a67-a470-9c0148b443e3", "external-id": "nsx-vlan-transportzone-784", "segmentation_id": 784, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56f47df0-50", "ovs_interfaceid": "56f47df0-50f1-43e0-a2d5-3a4ad9d19327", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 816.095838] env[62692]: DEBUG oslo_concurrency.lockutils [req-f4734622-b124-422c-9685-cbec71f22d0d req-c439e92c-bc06-4ddb-a154-9f47d2dbb7f8 service nova] Acquired lock "refresh_cache-aa0666ce-4d4f-4333-8acb-ea0cfb438cd9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.096062] env[62692]: DEBUG nova.network.neutron [req-f4734622-b124-422c-9685-cbec71f22d0d req-c439e92c-bc06-4ddb-a154-9f47d2dbb7f8 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Refreshing network info cache for port 56f47df0-50f1-43e0-a2d5-3a4ad9d19327 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 816.097325] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:98:ea:2a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd12aff80-9d1b-4a67-a470-9c0148b443e3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b0db2db5-00e9-4a3a-a41a-7ed7f4985073', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:b7:2e:8e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '233536d0-6913-4879-8442-42dcf1d4ecbb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '13349e97-1acd-48d2-a24b-44160a60773e', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:35:ae:c9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd12aff80-9d1b-4a67-a470-9c0148b443e3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '56f47df0-50f1-43e0-a2d5-3a4ad9d19327', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 816.109619] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Creating folder: Project (4fed583a514c4043a4b5772c73e21aff). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 816.114059] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ce92692e-f8e7-4bdf-b581-875fddd9e892 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.128990] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Created folder: Project (4fed583a514c4043a4b5772c73e21aff) in parent group-v248868. [ 816.129253] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Creating folder: Instances. Parent ref: group-v248916. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 816.129552] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-10030824-a4d9-4639-a5f9-3415c26708c6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.144546] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Created folder: Instances in parent group-v248916. [ 816.144831] env[62692]: DEBUG oslo.service.loopingcall [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.145040] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 816.145254] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-569c4d42-aae8-4a36-bb70-ba9adb1a187f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.178046] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 816.178046] env[62692]: value = "task-1141037" [ 816.178046] env[62692]: _type = "Task" [ 816.178046] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.181899] env[62692]: DEBUG oslo_vmware.api [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141034, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.190762] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141037, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.223169] env[62692]: INFO nova.compute.manager [-] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Took 1.43 seconds to deallocate network for instance. [ 816.401024] env[62692]: DEBUG nova.network.neutron [req-f4734622-b124-422c-9685-cbec71f22d0d req-c439e92c-bc06-4ddb-a154-9f47d2dbb7f8 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Updated VIF entry in instance network info cache for port 56f47df0-50f1-43e0-a2d5-3a4ad9d19327. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 816.401979] env[62692]: DEBUG nova.network.neutron [req-f4734622-b124-422c-9685-cbec71f22d0d req-c439e92c-bc06-4ddb-a154-9f47d2dbb7f8 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Updating instance_info_cache with network_info: [{"id": "b0db2db5-00e9-4a3a-a41a-7ed7f4985073", "address": "fa:16:3e:98:ea:2a", "network": {"id": "5f63a59b-8a80-45bb-8b7f-23bfc180fe98", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1827914724", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.125", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fed583a514c4043a4b5772c73e21aff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d12aff80-9d1b-4a67-a470-9c0148b443e3", "external-id": "nsx-vlan-transportzone-784", "segmentation_id": 784, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0db2db5-00", "ovs_interfaceid": "b0db2db5-00e9-4a3a-a41a-7ed7f4985073", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "13349e97-1acd-48d2-a24b-44160a60773e", "address": "fa:16:3e:b7:2e:8e", "network": {"id": "6ad813f0-48dd-45af-89d3-62d3ee8d084f", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-738939750", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.66", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4fed583a514c4043a4b5772c73e21aff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "233536d0-6913-4879-8442-42dcf1d4ecbb", "external-id": "nsx-vlan-transportzone-700", "segmentation_id": 700, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13349e97-1a", "ovs_interfaceid": "13349e97-1acd-48d2-a24b-44160a60773e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "56f47df0-50f1-43e0-a2d5-3a4ad9d19327", "address": "fa:16:3e:35:ae:c9", "network": {"id": "5f63a59b-8a80-45bb-8b7f-23bfc180fe98", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1827914724", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.174", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fed583a514c4043a4b5772c73e21aff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d12aff80-9d1b-4a67-a470-9c0148b443e3", "external-id": "nsx-vlan-transportzone-784", "segmentation_id": 784, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56f47df0-50", "ovs_interfaceid": "56f47df0-50f1-43e0-a2d5-3a4ad9d19327", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.412842] env[62692]: DEBUG oslo_vmware.api [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141033, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.598456} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.413763] env[62692]: INFO nova.virt.vmwareapi.ds_util [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64/81eb64fb-5a0c-47b7-9948-7dcb01b911bd-rescue.vmdk. [ 816.414661] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad7db0e0-33ef-4962-b994-c31032eb65de {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.442062] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Reconfiguring VM instance instance-0000002a to attach disk [datastore1] ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64/81eb64fb-5a0c-47b7-9948-7dcb01b911bd-rescue.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 816.442375] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5aa6167-d613-41f2-945b-dfc96bfc6997 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.463404] env[62692]: DEBUG oslo_vmware.api [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 816.463404] env[62692]: value = "task-1141038" [ 816.463404] env[62692]: _type = "Task" [ 816.463404] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.478232] env[62692]: DEBUG oslo_vmware.api [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141038, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.504671] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Acquiring lock "079ff57e-3733-4833-b376-0653e2a9af31" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.505115] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Lock "079ff57e-3733-4833-b376-0653e2a9af31" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.531036] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cbe8404-bb03-4178-8ff1-5d174f213ce8 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "27583cb4-0478-44d8-8970-478682271513" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 148.231s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.584334] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Acquiring lock "refresh_cache-81ce340c-fbef-4932-983c-595843530dbc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.584920] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Acquired lock "refresh_cache-81ce340c-fbef-4932-983c-595843530dbc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.584920] env[62692]: DEBUG nova.network.neutron [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 816.680897] env[62692]: DEBUG oslo_vmware.api [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141034, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.690546] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141037, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.729473] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.907147] env[62692]: DEBUG oslo_concurrency.lockutils [req-f4734622-b124-422c-9685-cbec71f22d0d req-c439e92c-bc06-4ddb-a154-9f47d2dbb7f8 service nova] Releasing lock "refresh_cache-aa0666ce-4d4f-4333-8acb-ea0cfb438cd9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.975290] env[62692]: DEBUG oslo_vmware.api [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141038, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.009673] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Lock "079ff57e-3733-4833-b376-0653e2a9af31" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.504s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.010453] env[62692]: DEBUG nova.compute.manager [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 817.035044] env[62692]: DEBUG nova.compute.manager [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 817.043369] env[62692]: DEBUG nova.compute.manager [req-7bee7d36-0e32-40b8-93e9-ddf6405925d0 req-1e4a452b-a9dc-474d-b44d-90273771c815 service nova] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Received event network-vif-deleted-e047ca79-8094-49de-809c-036fee17d2c5 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 817.122510] env[62692]: DEBUG nova.network.neutron [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 817.185542] env[62692]: DEBUG nova.compute.manager [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 817.185950] env[62692]: DEBUG nova.virt.hardware [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 817.186384] env[62692]: DEBUG nova.virt.hardware [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 817.186566] env[62692]: DEBUG nova.virt.hardware [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 817.186757] env[62692]: DEBUG nova.virt.hardware [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 817.186899] env[62692]: DEBUG nova.virt.hardware [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 817.187056] env[62692]: DEBUG nova.virt.hardware [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 817.187508] env[62692]: DEBUG nova.virt.hardware [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 817.187709] env[62692]: DEBUG nova.virt.hardware [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 817.187905] env[62692]: DEBUG nova.virt.hardware [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 817.188084] env[62692]: DEBUG nova.virt.hardware [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 817.188261] env[62692]: DEBUG nova.virt.hardware [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 817.190533] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e0e8c69-2acb-4739-b444-01739cb3afb2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.199170] env[62692]: DEBUG oslo_vmware.api [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141034, 'name': PowerOnVM_Task, 'duration_secs': 1.034247} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.200134] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 817.200372] env[62692]: INFO nova.compute.manager [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Took 6.49 seconds to spawn the instance on the hypervisor. [ 817.200664] env[62692]: DEBUG nova.compute.manager [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 817.204031] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc592275-7437-4a81-900b-79ee61d81115 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.212099] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141037, 'name': CreateVM_Task, 'duration_secs': 0.539217} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.213853] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-271d5097-cff9-4471-82a5-6f30f3ce0ceb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.219250] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 817.222338] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.222517] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.222855] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 817.230577] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb80c707-ea55-4e97-80aa-b2229393e10e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.254562] env[62692]: DEBUG oslo_vmware.api [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Waiting for the task: (returnval){ [ 817.254562] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52409881-a2a8-2505-6f0f-2551be74250a" [ 817.254562] env[62692]: _type = "Task" [ 817.254562] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.264106] env[62692]: DEBUG oslo_vmware.api [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52409881-a2a8-2505-6f0f-2551be74250a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.347142] env[62692]: DEBUG nova.network.neutron [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Updating instance_info_cache with network_info: [{"id": "6b631aba-53ff-466f-b003-cc92a2b87144", "address": "fa:16:3e:fc:74:b5", "network": {"id": "363f615c-5aea-4b08-83b0-6cc2588ff1a3", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1758578172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "198a5af1f19c49e8a3b070cfbe9ee01a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b631aba-53", "ovs_interfaceid": "6b631aba-53ff-466f-b003-cc92a2b87144", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.475427] env[62692]: DEBUG oslo_vmware.api [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141038, 'name': ReconfigVM_Task, 'duration_secs': 0.879127} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.478015] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Reconfigured VM instance instance-0000002a to attach disk [datastore1] ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64/81eb64fb-5a0c-47b7-9948-7dcb01b911bd-rescue.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 817.479062] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35017e92-8f40-4b8c-92d7-a5458abfd0b5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.508351] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-95359441-ad72-4963-a4ae-6d9cfd5e7383 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.520738] env[62692]: DEBUG nova.compute.utils [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 817.522255] env[62692]: DEBUG nova.compute.manager [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 817.522424] env[62692]: DEBUG nova.network.neutron [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 817.529959] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-253f4eda-ef40-4b0c-b1ef-a2523e8585d8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.534281] env[62692]: DEBUG oslo_vmware.api [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 817.534281] env[62692]: value = "task-1141039" [ 817.534281] env[62692]: _type = "Task" [ 817.534281] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.541908] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07fb07ca-19e8-424f-ae00-c4e7c15f07b5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.550767] env[62692]: DEBUG oslo_vmware.api [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141039, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.587156] env[62692]: DEBUG nova.policy [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d8f759b6dce4f6aabff7455bf422ec2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dab189f767684ddcbe16e49a051de941', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 817.589591] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.591156] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11944207-c201-4e59-969b-abfb7c0cd4f7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.602050] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c048c6b-4c27-4424-840e-0001b0751533 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.619566] env[62692]: DEBUG nova.compute.provider_tree [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 817.745618] env[62692]: INFO nova.compute.manager [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Took 21.20 seconds to build instance. [ 817.765402] env[62692]: DEBUG oslo_vmware.api [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52409881-a2a8-2505-6f0f-2551be74250a, 'name': SearchDatastore_Task, 'duration_secs': 0.013982} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.765692] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.766294] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 817.766294] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.766294] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.766545] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 817.766778] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-142985aa-2e69-4dcb-9d0a-68f23ea45426 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.777481] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 817.777653] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 817.780015] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-180b592c-dc8a-4cf8-aaec-672f155d8895 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.785096] env[62692]: DEBUG oslo_vmware.api [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Waiting for the task: (returnval){ [ 817.785096] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]520b41dd-5f51-36c8-f04a-c9811d75fcaf" [ 817.785096] env[62692]: _type = "Task" [ 817.785096] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.793517] env[62692]: DEBUG oslo_vmware.api [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]520b41dd-5f51-36c8-f04a-c9811d75fcaf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.853406] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Releasing lock "refresh_cache-81ce340c-fbef-4932-983c-595843530dbc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.853406] env[62692]: DEBUG nova.compute.manager [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Instance network_info: |[{"id": "6b631aba-53ff-466f-b003-cc92a2b87144", "address": "fa:16:3e:fc:74:b5", "network": {"id": "363f615c-5aea-4b08-83b0-6cc2588ff1a3", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1758578172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "198a5af1f19c49e8a3b070cfbe9ee01a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b631aba-53", "ovs_interfaceid": "6b631aba-53ff-466f-b003-cc92a2b87144", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 817.853832] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:74:b5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6b631aba-53ff-466f-b003-cc92a2b87144', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 817.861832] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Creating folder: Project (198a5af1f19c49e8a3b070cfbe9ee01a). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 817.862124] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-443c62c6-5999-4273-958b-f71b83f8b3e2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.871842] env[62692]: DEBUG nova.network.neutron [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Successfully created port: 60ca6789-edcf-48cb-855a-52e119e461e5 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 817.877346] env[62692]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 817.877521] env[62692]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62692) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 817.877926] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Folder already exists: Project (198a5af1f19c49e8a3b070cfbe9ee01a). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 817.878147] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Creating folder: Instances. Parent ref: group-v248872. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 817.878590] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e360fa97-fd97-4425-9248-7403b9af4c39 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.890414] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Created folder: Instances in parent group-v248872. [ 817.890692] env[62692]: DEBUG oslo.service.loopingcall [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 817.890959] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 817.891169] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2ce2a1c4-7b96-4971-9f46-842e906b4308 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.911829] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 817.911829] env[62692]: value = "task-1141042" [ 817.911829] env[62692]: _type = "Task" [ 817.911829] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.922226] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141042, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.026166] env[62692]: DEBUG nova.compute.manager [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 818.046324] env[62692]: DEBUG oslo_vmware.api [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141039, 'name': ReconfigVM_Task, 'duration_secs': 0.206104} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.047291] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 818.047291] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2f68fdba-d46d-4eb3-a264-6acb75fd18e7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.054598] env[62692]: DEBUG oslo_vmware.api [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 818.054598] env[62692]: value = "task-1141043" [ 818.054598] env[62692]: _type = "Task" [ 818.054598] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.066930] env[62692]: DEBUG oslo_vmware.api [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141043, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.092559] env[62692]: DEBUG nova.compute.manager [req-6eeb42a7-d69f-424a-9076-27183d05febb req-0099f3a5-2371-4064-9676-ce6d0c47b98d service nova] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Received event network-changed-6b631aba-53ff-466f-b003-cc92a2b87144 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 818.092837] env[62692]: DEBUG nova.compute.manager [req-6eeb42a7-d69f-424a-9076-27183d05febb req-0099f3a5-2371-4064-9676-ce6d0c47b98d service nova] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Refreshing instance network info cache due to event network-changed-6b631aba-53ff-466f-b003-cc92a2b87144. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 818.093174] env[62692]: DEBUG oslo_concurrency.lockutils [req-6eeb42a7-d69f-424a-9076-27183d05febb req-0099f3a5-2371-4064-9676-ce6d0c47b98d service nova] Acquiring lock "refresh_cache-81ce340c-fbef-4932-983c-595843530dbc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.093401] env[62692]: DEBUG oslo_concurrency.lockutils [req-6eeb42a7-d69f-424a-9076-27183d05febb req-0099f3a5-2371-4064-9676-ce6d0c47b98d service nova] Acquired lock "refresh_cache-81ce340c-fbef-4932-983c-595843530dbc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.093658] env[62692]: DEBUG nova.network.neutron [req-6eeb42a7-d69f-424a-9076-27183d05febb req-0099f3a5-2371-4064-9676-ce6d0c47b98d service nova] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Refreshing network info cache for port 6b631aba-53ff-466f-b003-cc92a2b87144 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 818.122460] env[62692]: DEBUG nova.scheduler.client.report [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 818.249014] env[62692]: DEBUG oslo_concurrency.lockutils [None req-65151137-3ca4-4a19-8001-262363c5b1b2 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Lock "1848c251-0a0c-4c36-8bd0-ff37befde2a0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 142.010s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.297464] env[62692]: DEBUG oslo_vmware.api [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]520b41dd-5f51-36c8-f04a-c9811d75fcaf, 'name': SearchDatastore_Task, 'duration_secs': 0.011026} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.298298] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfdc5900-99f4-456a-b3aa-4ce7f254880a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.305516] env[62692]: DEBUG oslo_vmware.api [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Waiting for the task: (returnval){ [ 818.305516] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5283b4c2-839a-ec51-f7f3-4bd51937827a" [ 818.305516] env[62692]: _type = "Task" [ 818.305516] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.314921] env[62692]: DEBUG oslo_vmware.api [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5283b4c2-839a-ec51-f7f3-4bd51937827a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.422600] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141042, 'name': CreateVM_Task, 'duration_secs': 0.345184} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.422819] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 818.423636] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'mount_device': '/dev/sda', 'attachment_id': '9a1ab797-e113-46a3-adbf-8248a4b72f20', 'guest_format': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-248875', 'volume_id': '9886d0c7-5df2-4bce-8853-faa1431b1865', 'name': 'volume-9886d0c7-5df2-4bce-8853-faa1431b1865', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '81ce340c-fbef-4932-983c-595843530dbc', 'attached_at': '', 'detached_at': '', 'volume_id': '9886d0c7-5df2-4bce-8853-faa1431b1865', 'serial': '9886d0c7-5df2-4bce-8853-faa1431b1865'}, 'device_type': None, 'delete_on_termination': True, 'disk_bus': None, 'boot_index': 0, 'volume_type': None}], 'swap': None} {{(pid=62692) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 818.423909] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Root volume attach. Driver type: vmdk {{(pid=62692) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 818.424819] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e43ed297-5b31-4b7d-9abc-c4de0b81cbfc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.434880] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f1d3ce-a9bb-4466-b372-ec10df5e8264 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.442142] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d104537d-7f1e-4664-b328-6d390c1fe172 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.450147] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-1a3f85a0-beb2-4f5c-b0e5-c73f3643814b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.459638] env[62692]: DEBUG oslo_vmware.api [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Waiting for the task: (returnval){ [ 818.459638] env[62692]: value = "task-1141044" [ 818.459638] env[62692]: _type = "Task" [ 818.459638] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.468028] env[62692]: DEBUG oslo_vmware.api [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Task: {'id': task-1141044, 'name': RelocateVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.565106] env[62692]: DEBUG oslo_vmware.api [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141043, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.628343] env[62692]: DEBUG oslo_concurrency.lockutils [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.630s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.629158] env[62692]: DEBUG nova.compute.manager [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 818.634354] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.199s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.635789] env[62692]: INFO nova.compute.claims [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 818.753568] env[62692]: DEBUG nova.compute.manager [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 818.819834] env[62692]: DEBUG oslo_vmware.api [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5283b4c2-839a-ec51-f7f3-4bd51937827a, 'name': SearchDatastore_Task, 'duration_secs': 0.010942} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.820342] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.820754] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] aa0666ce-4d4f-4333-8acb-ea0cfb438cd9/aa0666ce-4d4f-4333-8acb-ea0cfb438cd9.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 818.821220] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a5739146-5c76-4243-8c06-abd637a4833d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.829887] env[62692]: DEBUG oslo_vmware.api [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Waiting for the task: (returnval){ [ 818.829887] env[62692]: value = "task-1141045" [ 818.829887] env[62692]: _type = "Task" [ 818.829887] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.839669] env[62692]: DEBUG oslo_vmware.api [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141045, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.899153] env[62692]: DEBUG nova.network.neutron [req-6eeb42a7-d69f-424a-9076-27183d05febb req-0099f3a5-2371-4064-9676-ce6d0c47b98d service nova] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Updated VIF entry in instance network info cache for port 6b631aba-53ff-466f-b003-cc92a2b87144. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 818.899153] env[62692]: DEBUG nova.network.neutron [req-6eeb42a7-d69f-424a-9076-27183d05febb req-0099f3a5-2371-4064-9676-ce6d0c47b98d service nova] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Updating instance_info_cache with network_info: [{"id": "6b631aba-53ff-466f-b003-cc92a2b87144", "address": "fa:16:3e:fc:74:b5", "network": {"id": "363f615c-5aea-4b08-83b0-6cc2588ff1a3", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1758578172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "198a5af1f19c49e8a3b070cfbe9ee01a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b631aba-53", "ovs_interfaceid": "6b631aba-53ff-466f-b003-cc92a2b87144", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.971785] env[62692]: DEBUG oslo_vmware.api [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Task: {'id': task-1141044, 'name': RelocateVM_Task, 'duration_secs': 0.496977} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.972120] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Volume attach. Driver type: vmdk {{(pid=62692) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 818.972343] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-248875', 'volume_id': '9886d0c7-5df2-4bce-8853-faa1431b1865', 'name': 'volume-9886d0c7-5df2-4bce-8853-faa1431b1865', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '81ce340c-fbef-4932-983c-595843530dbc', 'attached_at': '', 'detached_at': '', 'volume_id': '9886d0c7-5df2-4bce-8853-faa1431b1865', 'serial': '9886d0c7-5df2-4bce-8853-faa1431b1865'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 818.973148] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35618398-2566-424f-ad12-4488575904bc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.990398] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4771e8e2-8827-46d3-93a9-a3d5e68ec8e4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.013985] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Reconfiguring VM instance instance-0000002f to attach disk [datastore2] volume-9886d0c7-5df2-4bce-8853-faa1431b1865/volume-9886d0c7-5df2-4bce-8853-faa1431b1865.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 819.014333] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ccde6876-716f-4d75-b70e-1d126bc96420 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.036399] env[62692]: DEBUG nova.compute.manager [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 819.038695] env[62692]: DEBUG oslo_vmware.api [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Waiting for the task: (returnval){ [ 819.038695] env[62692]: value = "task-1141046" [ 819.038695] env[62692]: _type = "Task" [ 819.038695] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.048667] env[62692]: DEBUG oslo_vmware.api [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Task: {'id': task-1141046, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.067578] env[62692]: DEBUG oslo_vmware.api [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141043, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.070115] env[62692]: DEBUG nova.virt.hardware [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 819.070480] env[62692]: DEBUG nova.virt.hardware [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 819.070584] env[62692]: DEBUG nova.virt.hardware [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 819.070767] env[62692]: DEBUG nova.virt.hardware [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 819.070920] env[62692]: DEBUG nova.virt.hardware [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 819.071126] env[62692]: DEBUG nova.virt.hardware [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 819.071356] env[62692]: DEBUG nova.virt.hardware [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 819.071515] env[62692]: DEBUG nova.virt.hardware [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 819.071686] env[62692]: DEBUG nova.virt.hardware [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 819.071854] env[62692]: DEBUG nova.virt.hardware [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 819.072039] env[62692]: DEBUG nova.virt.hardware [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 819.072969] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eea9148-5ae3-4bf9-8110-c1600c103dd0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.082277] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d33b1912-7720-4137-9aaf-c35ef91e4cc6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.142363] env[62692]: DEBUG nova.compute.utils [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 819.143865] env[62692]: DEBUG nova.compute.manager [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Not allocating networking since 'none' was specified. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 819.204722] env[62692]: INFO nova.compute.manager [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Rebuilding instance [ 819.254926] env[62692]: DEBUG nova.compute.manager [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 819.255871] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42ce5fc4-61b4-4b4c-a540-baf320c8e176 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.281677] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.343299] env[62692]: DEBUG oslo_vmware.api [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141045, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.503456} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.343581] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] aa0666ce-4d4f-4333-8acb-ea0cfb438cd9/aa0666ce-4d4f-4333-8acb-ea0cfb438cd9.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 819.343812] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 819.344094] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-11e1d5b7-78ca-4df3-9fe5-b4bb219a0e89 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.351239] env[62692]: DEBUG oslo_vmware.api [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Waiting for the task: (returnval){ [ 819.351239] env[62692]: value = "task-1141047" [ 819.351239] env[62692]: _type = "Task" [ 819.351239] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.364154] env[62692]: DEBUG oslo_vmware.api [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141047, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.401427] env[62692]: DEBUG oslo_concurrency.lockutils [req-6eeb42a7-d69f-424a-9076-27183d05febb req-0099f3a5-2371-4064-9676-ce6d0c47b98d service nova] Releasing lock "refresh_cache-81ce340c-fbef-4932-983c-595843530dbc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.401755] env[62692]: DEBUG nova.compute.manager [req-6eeb42a7-d69f-424a-9076-27183d05febb req-0099f3a5-2371-4064-9676-ce6d0c47b98d service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] Received event network-changed-d4991635-5c62-4473-9661-97ac75a187fb {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 819.402504] env[62692]: DEBUG nova.compute.manager [req-6eeb42a7-d69f-424a-9076-27183d05febb req-0099f3a5-2371-4064-9676-ce6d0c47b98d service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] Refreshing instance network info cache due to event network-changed-d4991635-5c62-4473-9661-97ac75a187fb. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 819.402787] env[62692]: DEBUG oslo_concurrency.lockutils [req-6eeb42a7-d69f-424a-9076-27183d05febb req-0099f3a5-2371-4064-9676-ce6d0c47b98d service nova] Acquiring lock "refresh_cache-27583cb4-0478-44d8-8970-478682271513" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.402935] env[62692]: DEBUG oslo_concurrency.lockutils [req-6eeb42a7-d69f-424a-9076-27183d05febb req-0099f3a5-2371-4064-9676-ce6d0c47b98d service nova] Acquired lock "refresh_cache-27583cb4-0478-44d8-8970-478682271513" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.403110] env[62692]: DEBUG nova.network.neutron [req-6eeb42a7-d69f-424a-9076-27183d05febb req-0099f3a5-2371-4064-9676-ce6d0c47b98d service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] Refreshing network info cache for port d4991635-5c62-4473-9661-97ac75a187fb {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 819.553935] env[62692]: DEBUG oslo_vmware.api [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Task: {'id': task-1141046, 'name': ReconfigVM_Task, 'duration_secs': 0.418678} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.554160] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Reconfigured VM instance instance-0000002f to attach disk [datastore2] volume-9886d0c7-5df2-4bce-8853-faa1431b1865/volume-9886d0c7-5df2-4bce-8853-faa1431b1865.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 819.563846] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a8b52537-4ed4-4b95-bc6e-83bcbf03dab7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.586597] env[62692]: DEBUG oslo_vmware.api [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141043, 'name': PowerOnVM_Task, 'duration_secs': 1.063247} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.588252] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 819.590241] env[62692]: DEBUG oslo_vmware.api [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Waiting for the task: (returnval){ [ 819.590241] env[62692]: value = "task-1141048" [ 819.590241] env[62692]: _type = "Task" [ 819.590241] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.591285] env[62692]: DEBUG nova.compute.manager [None req-3c891cdf-c23a-4f61-9e90-55fa8210a543 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 819.592185] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e72c16a0-76ac-4cde-9887-ad96b3ade0be {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.610082] env[62692]: DEBUG oslo_vmware.api [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Task: {'id': task-1141048, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.649979] env[62692]: DEBUG nova.compute.manager [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 819.773321] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 819.773321] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a7e3e731-3b3b-48d2-947f-7d84d279c715 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.781846] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Waiting for the task: (returnval){ [ 819.781846] env[62692]: value = "task-1141049" [ 819.781846] env[62692]: _type = "Task" [ 819.781846] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.801011] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141049, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.863106] env[62692]: DEBUG oslo_vmware.api [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141047, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066914} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.863459] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 819.864452] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1bda83e-a15d-4647-8af9-a3070d0e16d9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.895061] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Reconfiguring VM instance instance-0000002d to attach disk [datastore2] aa0666ce-4d4f-4333-8acb-ea0cfb438cd9/aa0666ce-4d4f-4333-8acb-ea0cfb438cd9.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 819.898824] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-00e776ed-b700-4bf4-b466-cc8c72290183 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.914480] env[62692]: DEBUG nova.compute.manager [req-c69fff35-cda9-4224-973f-bde0c0f34ea9 req-6d0af354-bb79-49c4-a81f-16123f93f9bf service nova] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Received event network-vif-plugged-60ca6789-edcf-48cb-855a-52e119e461e5 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 819.914760] env[62692]: DEBUG oslo_concurrency.lockutils [req-c69fff35-cda9-4224-973f-bde0c0f34ea9 req-6d0af354-bb79-49c4-a81f-16123f93f9bf service nova] Acquiring lock "1cdd1dbc-567d-46f6-9f05-dfbc57acbf99-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.914977] env[62692]: DEBUG oslo_concurrency.lockutils [req-c69fff35-cda9-4224-973f-bde0c0f34ea9 req-6d0af354-bb79-49c4-a81f-16123f93f9bf service nova] Lock "1cdd1dbc-567d-46f6-9f05-dfbc57acbf99-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.915173] env[62692]: DEBUG oslo_concurrency.lockutils [req-c69fff35-cda9-4224-973f-bde0c0f34ea9 req-6d0af354-bb79-49c4-a81f-16123f93f9bf service nova] Lock "1cdd1dbc-567d-46f6-9f05-dfbc57acbf99-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.915339] env[62692]: DEBUG nova.compute.manager [req-c69fff35-cda9-4224-973f-bde0c0f34ea9 req-6d0af354-bb79-49c4-a81f-16123f93f9bf service nova] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] No waiting events found dispatching network-vif-plugged-60ca6789-edcf-48cb-855a-52e119e461e5 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 819.915499] env[62692]: WARNING nova.compute.manager [req-c69fff35-cda9-4224-973f-bde0c0f34ea9 req-6d0af354-bb79-49c4-a81f-16123f93f9bf service nova] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Received unexpected event network-vif-plugged-60ca6789-edcf-48cb-855a-52e119e461e5 for instance with vm_state building and task_state spawning. [ 819.925815] env[62692]: DEBUG oslo_vmware.api [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Waiting for the task: (returnval){ [ 819.925815] env[62692]: value = "task-1141050" [ 819.925815] env[62692]: _type = "Task" [ 819.925815] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.937956] env[62692]: DEBUG oslo_vmware.api [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141050, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.105132] env[62692]: DEBUG oslo_vmware.api [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Task: {'id': task-1141048, 'name': ReconfigVM_Task, 'duration_secs': 0.336026} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.105132] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-248875', 'volume_id': '9886d0c7-5df2-4bce-8853-faa1431b1865', 'name': 'volume-9886d0c7-5df2-4bce-8853-faa1431b1865', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '81ce340c-fbef-4932-983c-595843530dbc', 'attached_at': '', 'detached_at': '', 'volume_id': '9886d0c7-5df2-4bce-8853-faa1431b1865', 'serial': '9886d0c7-5df2-4bce-8853-faa1431b1865'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 820.105709] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8af027b2-893b-4e7a-a15c-1afed7516de9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.118013] env[62692]: DEBUG oslo_vmware.api [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Waiting for the task: (returnval){ [ 820.118013] env[62692]: value = "task-1141051" [ 820.118013] env[62692]: _type = "Task" [ 820.118013] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.141144] env[62692]: DEBUG oslo_vmware.api [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Task: {'id': task-1141051, 'name': Rename_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.143826] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55aa94da-e74b-41f9-9f47-83342c4119ab {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.152327] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-544535cb-606e-4d0c-9794-89e374e6bd40 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.194202] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75df3f30-5c9a-4c3e-956a-ee2eb80b936b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.204022] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0117241c-8d72-4e88-b36e-7f687e418b17 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.219796] env[62692]: DEBUG nova.compute.provider_tree [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 820.234659] env[62692]: DEBUG nova.network.neutron [req-6eeb42a7-d69f-424a-9076-27183d05febb req-0099f3a5-2371-4064-9676-ce6d0c47b98d service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] Updated VIF entry in instance network info cache for port d4991635-5c62-4473-9661-97ac75a187fb. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 820.235129] env[62692]: DEBUG nova.network.neutron [req-6eeb42a7-d69f-424a-9076-27183d05febb req-0099f3a5-2371-4064-9676-ce6d0c47b98d service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] Updating instance_info_cache with network_info: [{"id": "d4991635-5c62-4473-9661-97ac75a187fb", "address": "fa:16:3e:d4:98:43", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4991635-5c", "ovs_interfaceid": "d4991635-5c62-4473-9661-97ac75a187fb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.293382] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141049, 'name': PowerOffVM_Task, 'duration_secs': 0.257733} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.293382] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 820.293744] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 820.294669] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-763f4e51-7be2-45f2-8f41-294116be47bc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.303017] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 820.303455] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6c51d2e7-b263-4dd3-bab9-b3bb19d3bd2b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.328562] env[62692]: DEBUG nova.network.neutron [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Successfully updated port: 60ca6789-edcf-48cb-855a-52e119e461e5 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 820.334992] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 820.334992] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 820.334992] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Deleting the datastore file [datastore2] 1848c251-0a0c-4c36-8bd0-ff37befde2a0 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 820.334992] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0b3b4583-395f-4439-8df7-261352c117f9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.342747] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Waiting for the task: (returnval){ [ 820.342747] env[62692]: value = "task-1141053" [ 820.342747] env[62692]: _type = "Task" [ 820.342747] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.352656] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141053, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.366137] env[62692]: DEBUG nova.compute.manager [req-e08f26b6-a3e0-4b87-a9f2-a71bca1bcb5d req-f6d07243-14c4-4856-8b7e-a3197eb87a3f service nova] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Received event network-changed-60ca6789-edcf-48cb-855a-52e119e461e5 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 820.366137] env[62692]: DEBUG nova.compute.manager [req-e08f26b6-a3e0-4b87-a9f2-a71bca1bcb5d req-f6d07243-14c4-4856-8b7e-a3197eb87a3f service nova] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Refreshing instance network info cache due to event network-changed-60ca6789-edcf-48cb-855a-52e119e461e5. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 820.366137] env[62692]: DEBUG oslo_concurrency.lockutils [req-e08f26b6-a3e0-4b87-a9f2-a71bca1bcb5d req-f6d07243-14c4-4856-8b7e-a3197eb87a3f service nova] Acquiring lock "refresh_cache-1cdd1dbc-567d-46f6-9f05-dfbc57acbf99" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.366359] env[62692]: DEBUG oslo_concurrency.lockutils [req-e08f26b6-a3e0-4b87-a9f2-a71bca1bcb5d req-f6d07243-14c4-4856-8b7e-a3197eb87a3f service nova] Acquired lock "refresh_cache-1cdd1dbc-567d-46f6-9f05-dfbc57acbf99" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.366578] env[62692]: DEBUG nova.network.neutron [req-e08f26b6-a3e0-4b87-a9f2-a71bca1bcb5d req-f6d07243-14c4-4856-8b7e-a3197eb87a3f service nova] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Refreshing network info cache for port 60ca6789-edcf-48cb-855a-52e119e461e5 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 820.437507] env[62692]: DEBUG oslo_vmware.api [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141050, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.640535] env[62692]: DEBUG oslo_vmware.api [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Task: {'id': task-1141051, 'name': Rename_Task, 'duration_secs': 0.363599} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.640835] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 820.641078] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-926e4ab2-b3c3-433a-821b-34b3aaf64eef {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.648264] env[62692]: DEBUG oslo_vmware.api [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Waiting for the task: (returnval){ [ 820.648264] env[62692]: value = "task-1141054" [ 820.648264] env[62692]: _type = "Task" [ 820.648264] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.657231] env[62692]: DEBUG oslo_vmware.api [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Task: {'id': task-1141054, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.665563] env[62692]: DEBUG nova.compute.manager [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 820.694097] env[62692]: DEBUG nova.virt.hardware [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 820.694394] env[62692]: DEBUG nova.virt.hardware [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 820.694550] env[62692]: DEBUG nova.virt.hardware [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 820.694728] env[62692]: DEBUG nova.virt.hardware [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 820.694873] env[62692]: DEBUG nova.virt.hardware [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 820.695032] env[62692]: DEBUG nova.virt.hardware [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 820.695657] env[62692]: DEBUG nova.virt.hardware [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 820.695914] env[62692]: DEBUG nova.virt.hardware [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 820.696234] env[62692]: DEBUG nova.virt.hardware [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 820.696527] env[62692]: DEBUG nova.virt.hardware [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 820.696818] env[62692]: DEBUG nova.virt.hardware [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 820.697984] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aee1829d-b975-46ae-9291-567cb7285aed {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.707875] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0677fdd-4e83-461e-ac29-054c4d4274f8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.722850] env[62692]: DEBUG nova.scheduler.client.report [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 820.726452] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Instance VIF info [] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 820.732251] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Creating folder: Project (11179a2ce977493da4c80874ce7454a3). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 820.732857] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c86b6831-cd24-48dc-b2c0-31a4d69da5fb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.737430] env[62692]: DEBUG oslo_concurrency.lockutils [req-6eeb42a7-d69f-424a-9076-27183d05febb req-0099f3a5-2371-4064-9676-ce6d0c47b98d service nova] Releasing lock "refresh_cache-27583cb4-0478-44d8-8970-478682271513" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.748357] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Created folder: Project (11179a2ce977493da4c80874ce7454a3) in parent group-v248868. [ 820.748716] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Creating folder: Instances. Parent ref: group-v248921. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 820.748880] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cef8b3f9-a6d1-406f-95b3-46868e67d006 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.766920] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Created folder: Instances in parent group-v248921. [ 820.767673] env[62692]: DEBUG oslo.service.loopingcall [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 820.767673] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 820.767847] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1040c460-cc43-42b9-abb7-98a7d8b8c3f1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.803764] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 820.803764] env[62692]: value = "task-1141057" [ 820.803764] env[62692]: _type = "Task" [ 820.803764] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.817643] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141057, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.831210] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Acquiring lock "refresh_cache-1cdd1dbc-567d-46f6-9f05-dfbc57acbf99" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.854872] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141053, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.113217} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.858192] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 820.858192] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 820.858192] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 820.916288] env[62692]: DEBUG nova.network.neutron [req-e08f26b6-a3e0-4b87-a9f2-a71bca1bcb5d req-f6d07243-14c4-4856-8b7e-a3197eb87a3f service nova] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 820.938180] env[62692]: DEBUG oslo_vmware.api [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141050, 'name': ReconfigVM_Task, 'duration_secs': 0.919756} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.938464] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Reconfigured VM instance instance-0000002d to attach disk [datastore2] aa0666ce-4d4f-4333-8acb-ea0cfb438cd9/aa0666ce-4d4f-4333-8acb-ea0cfb438cd9.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 820.939191] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4e6884ae-6fb9-4420-8e32-795d895bfb57 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.949751] env[62692]: DEBUG oslo_vmware.api [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Waiting for the task: (returnval){ [ 820.949751] env[62692]: value = "task-1141058" [ 820.949751] env[62692]: _type = "Task" [ 820.949751] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.959046] env[62692]: DEBUG oslo_vmware.api [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141058, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.032036] env[62692]: DEBUG nova.network.neutron [req-e08f26b6-a3e0-4b87-a9f2-a71bca1bcb5d req-f6d07243-14c4-4856-8b7e-a3197eb87a3f service nova] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.161873] env[62692]: DEBUG oslo_vmware.api [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Task: {'id': task-1141054, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.234584] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.600s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.235264] env[62692]: DEBUG nova.compute.manager [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 821.239239] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 12.429s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.239239] env[62692]: DEBUG nova.objects.instance [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62692) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 821.287303] env[62692]: DEBUG oslo_concurrency.lockutils [None req-725164a3-4b79-4d9a-846c-bf2588d13980 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "d99737d8-2eb0-40ee-b61e-6c736c84ea59" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.287607] env[62692]: DEBUG oslo_concurrency.lockutils [None req-725164a3-4b79-4d9a-846c-bf2588d13980 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "d99737d8-2eb0-40ee-b61e-6c736c84ea59" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.287893] env[62692]: DEBUG nova.compute.manager [None req-725164a3-4b79-4d9a-846c-bf2588d13980 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 821.289209] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7cd3b25-af7e-41ac-b43b-5a6192019d0a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.298080] env[62692]: DEBUG nova.compute.manager [None req-725164a3-4b79-4d9a-846c-bf2588d13980 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62692) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 821.298760] env[62692]: DEBUG nova.objects.instance [None req-725164a3-4b79-4d9a-846c-bf2588d13980 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lazy-loading 'flavor' on Instance uuid d99737d8-2eb0-40ee-b61e-6c736c84ea59 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 821.316675] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141057, 'name': CreateVM_Task, 'duration_secs': 0.344536} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.316896] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 821.318465] env[62692]: DEBUG oslo_concurrency.lockutils [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.322195] env[62692]: DEBUG oslo_concurrency.lockutils [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.322586] env[62692]: DEBUG oslo_concurrency.lockutils [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 821.323749] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6a79edb-1215-4ea2-8e89-bb8066e5e266 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.330533] env[62692]: DEBUG oslo_vmware.api [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Waiting for the task: (returnval){ [ 821.330533] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52e1c4e7-406c-9887-159b-672dbe3f51eb" [ 821.330533] env[62692]: _type = "Task" [ 821.330533] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.340909] env[62692]: DEBUG oslo_vmware.api [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52e1c4e7-406c-9887-159b-672dbe3f51eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.461530] env[62692]: DEBUG oslo_vmware.api [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141058, 'name': Rename_Task, 'duration_secs': 0.172373} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.461869] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 821.462147] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ad758cec-bce3-4520-b17b-54cfada0e74a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.470699] env[62692]: DEBUG oslo_vmware.api [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Waiting for the task: (returnval){ [ 821.470699] env[62692]: value = "task-1141059" [ 821.470699] env[62692]: _type = "Task" [ 821.470699] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.479461] env[62692]: DEBUG oslo_vmware.api [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141059, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.535106] env[62692]: DEBUG oslo_concurrency.lockutils [req-e08f26b6-a3e0-4b87-a9f2-a71bca1bcb5d req-f6d07243-14c4-4856-8b7e-a3197eb87a3f service nova] Releasing lock "refresh_cache-1cdd1dbc-567d-46f6-9f05-dfbc57acbf99" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.535545] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Acquired lock "refresh_cache-1cdd1dbc-567d-46f6-9f05-dfbc57acbf99" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.535720] env[62692]: DEBUG nova.network.neutron [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 821.660801] env[62692]: DEBUG oslo_vmware.api [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Task: {'id': task-1141054, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.746446] env[62692]: DEBUG nova.compute.utils [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 821.750575] env[62692]: DEBUG nova.compute.manager [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 821.751217] env[62692]: DEBUG nova.network.neutron [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 821.796891] env[62692]: DEBUG nova.policy [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a28f5e0e0af443c5be82214d446c221d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '589895ae20394e2fa525707bf7d09f2a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 821.807614] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-725164a3-4b79-4d9a-846c-bf2588d13980 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 821.807926] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-069d0691-ae38-45a5-ac7e-b24c60e19c0a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.816855] env[62692]: DEBUG oslo_vmware.api [None req-725164a3-4b79-4d9a-846c-bf2588d13980 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 821.816855] env[62692]: value = "task-1141060" [ 821.816855] env[62692]: _type = "Task" [ 821.816855] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.826378] env[62692]: DEBUG oslo_vmware.api [None req-725164a3-4b79-4d9a-846c-bf2588d13980 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141060, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.841276] env[62692]: DEBUG oslo_vmware.api [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52e1c4e7-406c-9887-159b-672dbe3f51eb, 'name': SearchDatastore_Task, 'duration_secs': 0.010496} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.841550] env[62692]: DEBUG oslo_concurrency.lockutils [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.841883] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 821.842071] env[62692]: DEBUG oslo_concurrency.lockutils [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.843023] env[62692]: DEBUG oslo_concurrency.lockutils [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.843023] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 821.843023] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4334f1e9-9caf-4c4d-b691-b7b7b6ccf0df {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.854537] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 821.854820] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 821.855617] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-46cc8467-378c-4ae7-b560-873ba67ca96f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.861959] env[62692]: DEBUG oslo_vmware.api [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Waiting for the task: (returnval){ [ 821.861959] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52e19669-c06a-b3e1-bdf7-fe11734862b4" [ 821.861959] env[62692]: _type = "Task" [ 821.861959] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.877809] env[62692]: DEBUG oslo_vmware.api [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52e19669-c06a-b3e1-bdf7-fe11734862b4, 'name': SearchDatastore_Task, 'duration_secs': 0.01133} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.878377] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3a465dc-10ae-47ac-a7ce-4d6c618c6e11 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.885845] env[62692]: DEBUG oslo_vmware.api [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Waiting for the task: (returnval){ [ 821.885845] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]526ef821-f463-ab27-37f1-f303ff8d9d7c" [ 821.885845] env[62692]: _type = "Task" [ 821.885845] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.903144] env[62692]: DEBUG oslo_vmware.api [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]526ef821-f463-ab27-37f1-f303ff8d9d7c, 'name': SearchDatastore_Task, 'duration_secs': 0.011921} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.905058] env[62692]: DEBUG nova.virt.hardware [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 821.907029] env[62692]: DEBUG nova.virt.hardware [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 821.907029] env[62692]: DEBUG nova.virt.hardware [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 821.907029] env[62692]: DEBUG nova.virt.hardware [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 821.907029] env[62692]: DEBUG nova.virt.hardware [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 821.907029] env[62692]: DEBUG nova.virt.hardware [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 821.907029] env[62692]: DEBUG nova.virt.hardware [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 821.907029] env[62692]: DEBUG nova.virt.hardware [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 821.907029] env[62692]: DEBUG nova.virt.hardware [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 821.907771] env[62692]: DEBUG nova.virt.hardware [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 821.908117] env[62692]: DEBUG nova.virt.hardware [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 821.911017] env[62692]: DEBUG oslo_concurrency.lockutils [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.911017] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 7a3f9c7d-03cf-4177-88db-acf33cc03feb/7a3f9c7d-03cf-4177-88db-acf33cc03feb.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 821.911017] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87b6dd26-0fad-4307-99a7-a33792a9d818 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.913194] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-036ba8d0-63eb-4532-98d2-21e3e8bacd4e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.922742] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-390c976f-658f-4458-9eae-c4aa98513306 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.929562] env[62692]: DEBUG oslo_vmware.api [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Waiting for the task: (returnval){ [ 821.929562] env[62692]: value = "task-1141061" [ 821.929562] env[62692]: _type = "Task" [ 821.929562] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.941578] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Instance VIF info [] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 821.947492] env[62692]: DEBUG oslo.service.loopingcall [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 821.948351] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 821.948564] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-89ed55c8-3535-463f-ac1e-35bc58e84879 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.966423] env[62692]: DEBUG oslo_vmware.api [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141061, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.976894] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 821.976894] env[62692]: value = "task-1141062" [ 821.976894] env[62692]: _type = "Task" [ 821.976894] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.983820] env[62692]: DEBUG oslo_vmware.api [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141059, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.989399] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141062, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.093897] env[62692]: DEBUG nova.network.neutron [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 822.167638] env[62692]: DEBUG oslo_vmware.api [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Task: {'id': task-1141054, 'name': PowerOnVM_Task, 'duration_secs': 1.458431} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.168343] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 822.168610] env[62692]: INFO nova.compute.manager [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Took 4.98 seconds to spawn the instance on the hypervisor. [ 822.168948] env[62692]: DEBUG nova.compute.manager [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 822.169988] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39b82a25-8a6c-40db-8609-b160e6bac0db {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.251803] env[62692]: DEBUG nova.compute.manager [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 822.255652] env[62692]: DEBUG nova.network.neutron [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Successfully created port: ce7100ab-22d6-4268-8f4c-844d9658c6d1 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 822.258490] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e921809d-95ae-4d9d-aae3-d0c7eb086144 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.020s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.260783] env[62692]: DEBUG oslo_concurrency.lockutils [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.403s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.262133] env[62692]: INFO nova.compute.claims [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 822.333030] env[62692]: DEBUG oslo_vmware.api [None req-725164a3-4b79-4d9a-846c-bf2588d13980 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141060, 'name': PowerOffVM_Task, 'duration_secs': 0.225162} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.333030] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-725164a3-4b79-4d9a-846c-bf2588d13980 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 822.333030] env[62692]: DEBUG nova.compute.manager [None req-725164a3-4b79-4d9a-846c-bf2588d13980 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 822.333030] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2b80c97-58bf-4342-8309-cc11ad4a204a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.443325] env[62692]: DEBUG oslo_vmware.api [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141061, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.488232] env[62692]: DEBUG oslo_vmware.api [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141059, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.492670] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141062, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.507224] env[62692]: DEBUG nova.network.neutron [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Updating instance_info_cache with network_info: [{"id": "60ca6789-edcf-48cb-855a-52e119e461e5", "address": "fa:16:3e:19:d3:63", "network": {"id": "79902a3d-8c64-454e-b5b1-dea4a3b960e5", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1711488481-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dab189f767684ddcbe16e49a051de941", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa410d21-2141-45bb-8d0b-16c77304605f", "external-id": "nsx-vlan-transportzone-886", "segmentation_id": 886, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60ca6789-ed", "ovs_interfaceid": "60ca6789-edcf-48cb-855a-52e119e461e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.695792] env[62692]: INFO nova.compute.manager [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Took 21.78 seconds to build instance. [ 822.851439] env[62692]: DEBUG oslo_concurrency.lockutils [None req-725164a3-4b79-4d9a-846c-bf2588d13980 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "d99737d8-2eb0-40ee-b61e-6c736c84ea59" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.563s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.937043] env[62692]: INFO nova.compute.manager [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Rescuing [ 822.937409] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "refresh_cache-bb547773-d176-4c8e-a0fa-a374d5050b1b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.937472] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquired lock "refresh_cache-bb547773-d176-4c8e-a0fa-a374d5050b1b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.937727] env[62692]: DEBUG nova.network.neutron [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 822.952959] env[62692]: DEBUG oslo_vmware.api [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141061, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.588786} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.954233] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 7a3f9c7d-03cf-4177-88db-acf33cc03feb/7a3f9c7d-03cf-4177-88db-acf33cc03feb.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 822.954776] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 822.955483] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8ce1e958-f680-41b1-aa8d-cfe0d65fc6af {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.964629] env[62692]: DEBUG oslo_vmware.api [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Waiting for the task: (returnval){ [ 822.964629] env[62692]: value = "task-1141063" [ 822.964629] env[62692]: _type = "Task" [ 822.964629] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.980566] env[62692]: DEBUG oslo_vmware.api [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141063, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.997708] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141062, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.998223] env[62692]: DEBUG oslo_vmware.api [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141059, 'name': PowerOnVM_Task, 'duration_secs': 1.3386} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.998673] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 822.999076] env[62692]: INFO nova.compute.manager [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Took 14.94 seconds to spawn the instance on the hypervisor. [ 822.999411] env[62692]: DEBUG nova.compute.manager [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 823.000341] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-025da694-ed03-4e1b-9000-345489ddab63 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.011658] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Releasing lock "refresh_cache-1cdd1dbc-567d-46f6-9f05-dfbc57acbf99" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.012231] env[62692]: DEBUG nova.compute.manager [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Instance network_info: |[{"id": "60ca6789-edcf-48cb-855a-52e119e461e5", "address": "fa:16:3e:19:d3:63", "network": {"id": "79902a3d-8c64-454e-b5b1-dea4a3b960e5", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1711488481-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dab189f767684ddcbe16e49a051de941", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa410d21-2141-45bb-8d0b-16c77304605f", "external-id": "nsx-vlan-transportzone-886", "segmentation_id": 886, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60ca6789-ed", "ovs_interfaceid": "60ca6789-edcf-48cb-855a-52e119e461e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 823.012863] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:19:d3:63', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aa410d21-2141-45bb-8d0b-16c77304605f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '60ca6789-edcf-48cb-855a-52e119e461e5', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 823.023798] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Creating folder: Project (dab189f767684ddcbe16e49a051de941). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 823.023798] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aaec63e1-dc61-4339-acee-3d450c1c2e72 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.034775] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Created folder: Project (dab189f767684ddcbe16e49a051de941) in parent group-v248868. [ 823.034775] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Creating folder: Instances. Parent ref: group-v248925. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 823.034775] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-25babb74-bf80-43af-884e-197d246bc7ff {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.046971] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Created folder: Instances in parent group-v248925. [ 823.048692] env[62692]: DEBUG oslo.service.loopingcall [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 823.048962] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 823.049242] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fb39555a-06da-41ba-bfa4-baa4a469f3f9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.074662] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 823.074662] env[62692]: value = "task-1141066" [ 823.074662] env[62692]: _type = "Task" [ 823.074662] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.085124] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141066, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.197742] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c4126f19-5b18-4ca9-ae12-8098e9f150a5 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Lock "81ce340c-fbef-4932-983c-595843530dbc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.837s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.269746] env[62692]: DEBUG nova.compute.manager [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 823.301711] env[62692]: DEBUG nova.virt.hardware [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 823.301973] env[62692]: DEBUG nova.virt.hardware [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 823.302161] env[62692]: DEBUG nova.virt.hardware [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 823.302355] env[62692]: DEBUG nova.virt.hardware [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 823.302546] env[62692]: DEBUG nova.virt.hardware [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 823.302639] env[62692]: DEBUG nova.virt.hardware [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 823.302877] env[62692]: DEBUG nova.virt.hardware [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 823.303397] env[62692]: DEBUG nova.virt.hardware [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 823.303863] env[62692]: DEBUG nova.virt.hardware [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 823.304206] env[62692]: DEBUG nova.virt.hardware [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 823.304583] env[62692]: DEBUG nova.virt.hardware [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 823.305821] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c7f5a98-51b8-4ce7-9e36-cda91f362a55 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.323233] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16f43e1b-c6c2-4497-aa54-46e6a0ea3a64 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.477066] env[62692]: DEBUG oslo_vmware.api [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141063, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089183} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.477634] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 823.480167] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-691c4a8f-2936-4747-b84a-2b4421f9e4de {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.497680] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141062, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.515091] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] 7a3f9c7d-03cf-4177-88db-acf33cc03feb/7a3f9c7d-03cf-4177-88db-acf33cc03feb.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 823.522350] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7fd241e6-3bee-48c3-8234-283f6b427126 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.547157] env[62692]: INFO nova.compute.manager [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Took 27.05 seconds to build instance. [ 823.554252] env[62692]: DEBUG nova.compute.manager [req-d499de2c-6138-4362-8492-e84793952e74 req-4aed7556-b9c6-44fe-a5a9-3eebef56c690 service nova] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Received event network-changed-6b631aba-53ff-466f-b003-cc92a2b87144 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 823.558020] env[62692]: DEBUG nova.compute.manager [req-d499de2c-6138-4362-8492-e84793952e74 req-4aed7556-b9c6-44fe-a5a9-3eebef56c690 service nova] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Refreshing instance network info cache due to event network-changed-6b631aba-53ff-466f-b003-cc92a2b87144. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 823.558020] env[62692]: DEBUG oslo_concurrency.lockutils [req-d499de2c-6138-4362-8492-e84793952e74 req-4aed7556-b9c6-44fe-a5a9-3eebef56c690 service nova] Acquiring lock "refresh_cache-81ce340c-fbef-4932-983c-595843530dbc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.558020] env[62692]: DEBUG oslo_concurrency.lockutils [req-d499de2c-6138-4362-8492-e84793952e74 req-4aed7556-b9c6-44fe-a5a9-3eebef56c690 service nova] Acquired lock "refresh_cache-81ce340c-fbef-4932-983c-595843530dbc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.558020] env[62692]: DEBUG nova.network.neutron [req-d499de2c-6138-4362-8492-e84793952e74 req-4aed7556-b9c6-44fe-a5a9-3eebef56c690 service nova] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Refreshing network info cache for port 6b631aba-53ff-466f-b003-cc92a2b87144 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 823.560250] env[62692]: DEBUG oslo_vmware.api [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Waiting for the task: (returnval){ [ 823.560250] env[62692]: value = "task-1141067" [ 823.560250] env[62692]: _type = "Task" [ 823.560250] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.588543] env[62692]: DEBUG oslo_vmware.api [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141067, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.598941] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141066, 'name': CreateVM_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.701737] env[62692]: DEBUG nova.compute.manager [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 823.808435] env[62692]: DEBUG nova.network.neutron [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Updating instance_info_cache with network_info: [{"id": "c77af962-553d-4504-825e-78a6689aee79", "address": "fa:16:3e:9a:01:cc", "network": {"id": "17e72c3b-603f-4305-96ac-89a9b4c9c7e9", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-123501993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7db5465cb5db465b92dfa570572b6e80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc77af962-55", "ovs_interfaceid": "c77af962-553d-4504-825e-78a6689aee79", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.835386] env[62692]: DEBUG nova.objects.instance [None req-6240b800-2ca6-4456-90a7-7097659c4b52 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lazy-loading 'flavor' on Instance uuid d99737d8-2eb0-40ee-b61e-6c736c84ea59 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 823.842423] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbd4e06d-88fa-41b5-9f26-6ed6d774494a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.851053] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f704317-216b-4e74-823a-99e368386c57 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.907971] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8342170f-d714-4769-8a8e-340c35e0199c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.920187] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1faa53c3-c879-4f19-96bf-6454c74af6bf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.939614] env[62692]: DEBUG nova.compute.provider_tree [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 824.008706] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141062, 'name': CreateVM_Task, 'duration_secs': 1.612896} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.008937] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 824.009559] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.009861] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.010324] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 824.011576] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-762acf72-9070-4f34-9a43-0b4710e09852 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.019042] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Waiting for the task: (returnval){ [ 824.019042] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]520f2c8e-63b7-a6e3-5685-6a942b1e00ac" [ 824.019042] env[62692]: _type = "Task" [ 824.019042] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.030438] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]520f2c8e-63b7-a6e3-5685-6a942b1e00ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.053902] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bf198b0e-d4a7-479c-badc-cd56247cd0b9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Lock "aa0666ce-4d4f-4333-8acb-ea0cfb438cd9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.606s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.078858] env[62692]: DEBUG nova.compute.manager [req-3900c210-e5a3-4fcc-95a1-d415eed1ed17 req-79a84b51-56bf-434a-a7cd-7ea136bda65e service nova] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Received event network-vif-plugged-ce7100ab-22d6-4268-8f4c-844d9658c6d1 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 824.080437] env[62692]: DEBUG oslo_concurrency.lockutils [req-3900c210-e5a3-4fcc-95a1-d415eed1ed17 req-79a84b51-56bf-434a-a7cd-7ea136bda65e service nova] Acquiring lock "b1d6409a-2733-470f-a929-672fe1631b1b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.080437] env[62692]: DEBUG oslo_concurrency.lockutils [req-3900c210-e5a3-4fcc-95a1-d415eed1ed17 req-79a84b51-56bf-434a-a7cd-7ea136bda65e service nova] Lock "b1d6409a-2733-470f-a929-672fe1631b1b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.080437] env[62692]: DEBUG oslo_concurrency.lockutils [req-3900c210-e5a3-4fcc-95a1-d415eed1ed17 req-79a84b51-56bf-434a-a7cd-7ea136bda65e service nova] Lock "b1d6409a-2733-470f-a929-672fe1631b1b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.080437] env[62692]: DEBUG nova.compute.manager [req-3900c210-e5a3-4fcc-95a1-d415eed1ed17 req-79a84b51-56bf-434a-a7cd-7ea136bda65e service nova] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] No waiting events found dispatching network-vif-plugged-ce7100ab-22d6-4268-8f4c-844d9658c6d1 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 824.082070] env[62692]: WARNING nova.compute.manager [req-3900c210-e5a3-4fcc-95a1-d415eed1ed17 req-79a84b51-56bf-434a-a7cd-7ea136bda65e service nova] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Received unexpected event network-vif-plugged-ce7100ab-22d6-4268-8f4c-844d9658c6d1 for instance with vm_state building and task_state spawning. [ 824.109156] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141066, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.109515] env[62692]: DEBUG oslo_vmware.api [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141067, 'name': ReconfigVM_Task, 'duration_secs': 0.402659} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.109788] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Reconfigured VM instance instance-00000031 to attach disk [datastore1] 7a3f9c7d-03cf-4177-88db-acf33cc03feb/7a3f9c7d-03cf-4177-88db-acf33cc03feb.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 824.111069] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-92b6d59c-16d5-414c-b110-5a1f00463e86 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.120458] env[62692]: DEBUG oslo_vmware.api [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Waiting for the task: (returnval){ [ 824.120458] env[62692]: value = "task-1141068" [ 824.120458] env[62692]: _type = "Task" [ 824.120458] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.131531] env[62692]: DEBUG oslo_vmware.api [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141068, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.235308] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.311191] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Releasing lock "refresh_cache-bb547773-d176-4c8e-a0fa-a374d5050b1b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.345281] env[62692]: DEBUG nova.network.neutron [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Successfully updated port: ce7100ab-22d6-4268-8f4c-844d9658c6d1 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 824.345281] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6240b800-2ca6-4456-90a7-7097659c4b52 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "refresh_cache-d99737d8-2eb0-40ee-b61e-6c736c84ea59" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.346177] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6240b800-2ca6-4456-90a7-7097659c4b52 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired lock "refresh_cache-d99737d8-2eb0-40ee-b61e-6c736c84ea59" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.346469] env[62692]: DEBUG nova.network.neutron [None req-6240b800-2ca6-4456-90a7-7097659c4b52 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 824.346636] env[62692]: DEBUG nova.objects.instance [None req-6240b800-2ca6-4456-90a7-7097659c4b52 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lazy-loading 'info_cache' on Instance uuid d99737d8-2eb0-40ee-b61e-6c736c84ea59 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 824.442959] env[62692]: DEBUG nova.scheduler.client.report [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 824.524888] env[62692]: DEBUG nova.network.neutron [req-d499de2c-6138-4362-8492-e84793952e74 req-4aed7556-b9c6-44fe-a5a9-3eebef56c690 service nova] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Updated VIF entry in instance network info cache for port 6b631aba-53ff-466f-b003-cc92a2b87144. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 824.525414] env[62692]: DEBUG nova.network.neutron [req-d499de2c-6138-4362-8492-e84793952e74 req-4aed7556-b9c6-44fe-a5a9-3eebef56c690 service nova] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Updating instance_info_cache with network_info: [{"id": "6b631aba-53ff-466f-b003-cc92a2b87144", "address": "fa:16:3e:fc:74:b5", "network": {"id": "363f615c-5aea-4b08-83b0-6cc2588ff1a3", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1758578172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "198a5af1f19c49e8a3b070cfbe9ee01a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b631aba-53", "ovs_interfaceid": "6b631aba-53ff-466f-b003-cc92a2b87144", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.534744] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]520f2c8e-63b7-a6e3-5685-6a942b1e00ac, 'name': SearchDatastore_Task, 'duration_secs': 0.01193} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.535193] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.535448] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 824.535680] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.536320] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.536320] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 824.536320] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-49868b58-54fb-4cd1-ab07-038b3bd7521e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.547311] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 824.547311] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 824.548133] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e73c3c36-82a8-4d78-803f-bfcd1fe97699 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.555736] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Waiting for the task: (returnval){ [ 824.555736] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52e543d5-08b5-7576-4f32-ff182d2ac5bc" [ 824.555736] env[62692]: _type = "Task" [ 824.555736] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.560889] env[62692]: DEBUG nova.compute.manager [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 824.567200] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52e543d5-08b5-7576-4f32-ff182d2ac5bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.596490] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141066, 'name': CreateVM_Task, 'duration_secs': 1.106245} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.600020] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 824.600020] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.600020] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.600020] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 824.600020] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a8fc52c-e642-448f-9340-ebd3623fdb2d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.606032] env[62692]: DEBUG oslo_vmware.api [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Waiting for the task: (returnval){ [ 824.606032] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]529060c0-253b-68f7-4007-5ec803776cb0" [ 824.606032] env[62692]: _type = "Task" [ 824.606032] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.624694] env[62692]: DEBUG oslo_vmware.api [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]529060c0-253b-68f7-4007-5ec803776cb0, 'name': SearchDatastore_Task, 'duration_secs': 0.011155} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.627825] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.628077] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 824.628305] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.628450] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.628625] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 824.629161] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c2568928-c780-40d0-af13-7a9c8cfd3bab {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.636440] env[62692]: DEBUG oslo_vmware.api [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141068, 'name': Rename_Task, 'duration_secs': 0.166538} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.636843] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 824.636946] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6a0bb020-bc49-465a-a8fc-180a982f8176 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.639602] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 824.639798] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 824.641214] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2dc83c8c-e52d-47b6-9b6e-928db31410f9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.646280] env[62692]: DEBUG oslo_vmware.api [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Waiting for the task: (returnval){ [ 824.646280] env[62692]: value = "task-1141069" [ 824.646280] env[62692]: _type = "Task" [ 824.646280] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.647502] env[62692]: DEBUG oslo_vmware.api [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Waiting for the task: (returnval){ [ 824.647502] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]521e315c-5a94-6721-1bbe-e9ef71b65682" [ 824.647502] env[62692]: _type = "Task" [ 824.647502] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.662884] env[62692]: DEBUG oslo_vmware.api [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141069, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.663100] env[62692]: DEBUG oslo_vmware.api [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]521e315c-5a94-6721-1bbe-e9ef71b65682, 'name': SearchDatastore_Task, 'duration_secs': 0.010916} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.663828] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd78b024-3913-4fcb-9df7-4c19e614fc67 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.669777] env[62692]: DEBUG oslo_vmware.api [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Waiting for the task: (returnval){ [ 824.669777] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52add908-16e1-bbdf-b644-04eb3226ab33" [ 824.669777] env[62692]: _type = "Task" [ 824.669777] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.682155] env[62692]: DEBUG oslo_vmware.api [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52add908-16e1-bbdf-b644-04eb3226ab33, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.850084] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "refresh_cache-b1d6409a-2733-470f-a929-672fe1631b1b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.850328] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired lock "refresh_cache-b1d6409a-2733-470f-a929-672fe1631b1b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.850604] env[62692]: DEBUG nova.network.neutron [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 824.853451] env[62692]: DEBUG nova.objects.base [None req-6240b800-2ca6-4456-90a7-7097659c4b52 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62692) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 824.860630] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 824.860630] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-23ec0c1e-7f7b-4871-9877-5283e6a8bda2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.872322] env[62692]: DEBUG oslo_vmware.api [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 824.872322] env[62692]: value = "task-1141070" [ 824.872322] env[62692]: _type = "Task" [ 824.872322] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.882833] env[62692]: DEBUG oslo_vmware.api [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141070, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.947877] env[62692]: DEBUG oslo_concurrency.lockutils [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.687s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.948448] env[62692]: DEBUG nova.compute.manager [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 824.951741] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.095s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.952107] env[62692]: DEBUG nova.objects.instance [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Lazy-loading 'resources' on Instance uuid 105e6fd5-4eff-4a0c-9a4b-e6deade781d5 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 825.030151] env[62692]: DEBUG oslo_concurrency.lockutils [req-d499de2c-6138-4362-8492-e84793952e74 req-4aed7556-b9c6-44fe-a5a9-3eebef56c690 service nova] Releasing lock "refresh_cache-81ce340c-fbef-4932-983c-595843530dbc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.073660] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52e543d5-08b5-7576-4f32-ff182d2ac5bc, 'name': SearchDatastore_Task, 'duration_secs': 0.010843} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.077778] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e2eb3d6-6c83-46b7-a5d5-54bcfc29afce {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.084916] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Waiting for the task: (returnval){ [ 825.084916] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52d7adf2-ae03-c02d-8d19-3ceae2730999" [ 825.084916] env[62692]: _type = "Task" [ 825.084916] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.098827] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52d7adf2-ae03-c02d-8d19-3ceae2730999, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.099678] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.158474] env[62692]: DEBUG oslo_vmware.api [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141069, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.178090] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Acquiring lock "aa0666ce-4d4f-4333-8acb-ea0cfb438cd9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.178371] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Lock "aa0666ce-4d4f-4333-8acb-ea0cfb438cd9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.178696] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Acquiring lock "aa0666ce-4d4f-4333-8acb-ea0cfb438cd9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.178909] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Lock "aa0666ce-4d4f-4333-8acb-ea0cfb438cd9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.179096] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Lock "aa0666ce-4d4f-4333-8acb-ea0cfb438cd9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.184195] env[62692]: DEBUG oslo_vmware.api [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52add908-16e1-bbdf-b644-04eb3226ab33, 'name': SearchDatastore_Task, 'duration_secs': 0.022498} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.184639] env[62692]: INFO nova.compute.manager [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Terminating instance [ 825.186213] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.186461] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99/1cdd1dbc-567d-46f6-9f05-dfbc57acbf99.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 825.187076] env[62692]: DEBUG nova.compute.manager [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 825.187268] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 825.187512] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ccae432f-3610-4459-aba8-0f358106ec91 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.190084] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87380c94-83e6-44f8-b3a9-458e1ea1a6d6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.198450] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 825.199817] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-50bd5563-1ab8-4072-b01b-ed8692fe5efc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.201475] env[62692]: DEBUG oslo_vmware.api [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Waiting for the task: (returnval){ [ 825.201475] env[62692]: value = "task-1141071" [ 825.201475] env[62692]: _type = "Task" [ 825.201475] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.208744] env[62692]: DEBUG oslo_vmware.api [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Waiting for the task: (returnval){ [ 825.208744] env[62692]: value = "task-1141072" [ 825.208744] env[62692]: _type = "Task" [ 825.208744] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.212132] env[62692]: DEBUG oslo_vmware.api [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Task: {'id': task-1141071, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.220914] env[62692]: DEBUG oslo_vmware.api [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141072, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.384925] env[62692]: DEBUG oslo_vmware.api [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141070, 'name': PowerOffVM_Task, 'duration_secs': 0.248994} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.385219] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 825.386044] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bbccd92-c954-4ef5-ae83-9e802cd6d71d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.407200] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d0c3246-af41-401c-a36e-75fb42d5428b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.421961] env[62692]: DEBUG nova.network.neutron [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 825.453441] env[62692]: DEBUG nova.compute.utils [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 825.457564] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 825.461140] env[62692]: DEBUG nova.compute.manager [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 825.461409] env[62692]: DEBUG nova.network.neutron [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 825.467275] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fa3bc837-4cd4-4f9d-bab9-e5007cd3f547 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.477053] env[62692]: DEBUG oslo_vmware.api [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 825.477053] env[62692]: value = "task-1141073" [ 825.477053] env[62692]: _type = "Task" [ 825.477053] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.495115] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] VM already powered off {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 825.495336] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 825.495574] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.495714] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.495887] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 825.500023] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d51f4a31-fa4f-4aa3-ac4a-dbdd7818742f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.514919] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 825.515134] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 825.516023] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2efc12c0-bb64-4609-9e9d-70dd56879927 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.527811] env[62692]: DEBUG oslo_vmware.api [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 825.527811] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52f584c1-7b02-f444-693b-90f4a2696560" [ 825.527811] env[62692]: _type = "Task" [ 825.527811] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.537964] env[62692]: DEBUG oslo_vmware.api [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52f584c1-7b02-f444-693b-90f4a2696560, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.550359] env[62692]: DEBUG nova.policy [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '55ba5fcf4b7e47b1aab816017f4bcabf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '454fce691f174c7e9348638d98597a26', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 825.603636] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52d7adf2-ae03-c02d-8d19-3ceae2730999, 'name': SearchDatastore_Task, 'duration_secs': 0.03151} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.603636] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.603636] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 1848c251-0a0c-4c36-8bd0-ff37befde2a0/1848c251-0a0c-4c36-8bd0-ff37befde2a0.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 825.603636] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c1224751-d7d2-47e4-a27e-944d615c9402 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.618261] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Waiting for the task: (returnval){ [ 825.618261] env[62692]: value = "task-1141074" [ 825.618261] env[62692]: _type = "Task" [ 825.618261] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.632425] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141074, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.659706] env[62692]: DEBUG oslo_vmware.api [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141069, 'name': PowerOnVM_Task, 'duration_secs': 0.81407} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.660084] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 825.660676] env[62692]: INFO nova.compute.manager [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Took 4.99 seconds to spawn the instance on the hypervisor. [ 825.660676] env[62692]: DEBUG nova.compute.manager [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 825.664114] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cd4f45c-e992-4e1d-844e-a26cd20d6022 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.719972] env[62692]: DEBUG oslo_vmware.api [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Task: {'id': task-1141071, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.728166] env[62692]: DEBUG oslo_vmware.api [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141072, 'name': PowerOffVM_Task, 'duration_secs': 0.26584} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.728166] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 825.728166] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 825.728639] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d887fe02-f50a-41d0-b942-86e1fa4e234f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.782673] env[62692]: DEBUG nova.network.neutron [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Updating instance_info_cache with network_info: [{"id": "ce7100ab-22d6-4268-8f4c-844d9658c6d1", "address": "fa:16:3e:59:21:56", "network": {"id": "dd260020-2d74-46ba-aa78-af29bdefa176", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-617179655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "589895ae20394e2fa525707bf7d09f2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce7100ab-22", "ovs_interfaceid": "ce7100ab-22d6-4268-8f4c-844d9658c6d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.853465] env[62692]: DEBUG nova.network.neutron [None req-6240b800-2ca6-4456-90a7-7097659c4b52 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Updating instance_info_cache with network_info: [{"id": "7f2cabdf-cbda-4cf7-a518-13f619716d12", "address": "fa:16:3e:6b:22:0c", "network": {"id": "f70727d6-a9aa-44b0-a244-2533b3f96571", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-284105837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1037e5d436bd429391159b87a712f6e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f2cabdf-cb", "ovs_interfaceid": "7f2cabdf-cbda-4cf7-a518-13f619716d12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.960024] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 825.960024] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 825.960024] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Deleting the datastore file [datastore2] aa0666ce-4d4f-4333-8acb-ea0cfb438cd9 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 825.960024] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-481d53ed-7a81-469a-8ab6-50442fd55f1b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.972164] env[62692]: DEBUG nova.compute.manager [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 825.981978] env[62692]: DEBUG oslo_vmware.api [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Waiting for the task: (returnval){ [ 825.981978] env[62692]: value = "task-1141076" [ 825.981978] env[62692]: _type = "Task" [ 825.981978] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.992026] env[62692]: DEBUG oslo_vmware.api [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141076, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.017243] env[62692]: DEBUG nova.network.neutron [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Successfully created port: d996a5d5-fb43-4af0-9531-397082ddca3f {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 826.054280] env[62692]: DEBUG oslo_vmware.api [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52f584c1-7b02-f444-693b-90f4a2696560, 'name': SearchDatastore_Task, 'duration_secs': 0.068649} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.056186] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0fb12080-9cbe-44e1-82f3-b9512b56f99e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.066284] env[62692]: DEBUG oslo_vmware.api [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 826.066284] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52a38641-0524-66dc-d7cc-ade63a50539d" [ 826.066284] env[62692]: _type = "Task" [ 826.066284] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.068157] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf86fdab-594c-44a9-b09a-ad53b2dfff0d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.086152] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67712dda-ea1e-4c11-bb12-7fa3f92bab76 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.089968] env[62692]: DEBUG oslo_vmware.api [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52a38641-0524-66dc-d7cc-ade63a50539d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.125041] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14b90020-848b-448e-9e16-21bd1ab7c019 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.135025] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141074, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.139009] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b89c8d8-4bd9-42f6-9450-ec69e9c3d488 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.155834] env[62692]: DEBUG nova.compute.provider_tree [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 826.192482] env[62692]: INFO nova.compute.manager [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Took 23.81 seconds to build instance. [ 826.215217] env[62692]: DEBUG oslo_vmware.api [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Task: {'id': task-1141071, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.654615} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.215490] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99/1cdd1dbc-567d-46f6-9f05-dfbc57acbf99.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 826.215797] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 826.216134] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0349e90d-2cae-4e18-a97b-4e3b560052ad {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.228923] env[62692]: DEBUG oslo_vmware.api [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Waiting for the task: (returnval){ [ 826.228923] env[62692]: value = "task-1141077" [ 826.228923] env[62692]: _type = "Task" [ 826.228923] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.241576] env[62692]: DEBUG oslo_vmware.api [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Task: {'id': task-1141077, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.287728] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Releasing lock "refresh_cache-b1d6409a-2733-470f-a929-672fe1631b1b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.288175] env[62692]: DEBUG nova.compute.manager [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Instance network_info: |[{"id": "ce7100ab-22d6-4268-8f4c-844d9658c6d1", "address": "fa:16:3e:59:21:56", "network": {"id": "dd260020-2d74-46ba-aa78-af29bdefa176", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-617179655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "589895ae20394e2fa525707bf7d09f2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce7100ab-22", "ovs_interfaceid": "ce7100ab-22d6-4268-8f4c-844d9658c6d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 826.288830] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:59:21:56', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55764410-260e-4339-a020-6b30995584bf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ce7100ab-22d6-4268-8f4c-844d9658c6d1', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 826.301243] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Creating folder: Project (589895ae20394e2fa525707bf7d09f2a). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 826.301650] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5df442e8-ba10-4dbb-abd1-9bbcbda8f064 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.321157] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Created folder: Project (589895ae20394e2fa525707bf7d09f2a) in parent group-v248868. [ 826.321426] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Creating folder: Instances. Parent ref: group-v248928. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 826.321745] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d733babb-f42a-4b38-a477-381e281be7cb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.339916] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Created folder: Instances in parent group-v248928. [ 826.340346] env[62692]: DEBUG oslo.service.loopingcall [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 826.340633] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 826.340996] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6b224d34-820a-4341-b983-f022e89aeaef {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.361280] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6240b800-2ca6-4456-90a7-7097659c4b52 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Releasing lock "refresh_cache-d99737d8-2eb0-40ee-b61e-6c736c84ea59" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.373372] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 826.373372] env[62692]: value = "task-1141080" [ 826.373372] env[62692]: _type = "Task" [ 826.373372] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.386536] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141080, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.508707] env[62692]: DEBUG oslo_vmware.api [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141076, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.531629] env[62692]: DEBUG nova.compute.manager [req-d982430d-4383-46d1-a0e0-6687f032c977 req-d4ab2b5c-3973-41cc-b682-ffca7c702429 service nova] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Received event network-changed-ce7100ab-22d6-4268-8f4c-844d9658c6d1 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 826.531868] env[62692]: DEBUG nova.compute.manager [req-d982430d-4383-46d1-a0e0-6687f032c977 req-d4ab2b5c-3973-41cc-b682-ffca7c702429 service nova] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Refreshing instance network info cache due to event network-changed-ce7100ab-22d6-4268-8f4c-844d9658c6d1. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 826.534305] env[62692]: DEBUG oslo_concurrency.lockutils [req-d982430d-4383-46d1-a0e0-6687f032c977 req-d4ab2b5c-3973-41cc-b682-ffca7c702429 service nova] Acquiring lock "refresh_cache-b1d6409a-2733-470f-a929-672fe1631b1b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.534473] env[62692]: DEBUG oslo_concurrency.lockutils [req-d982430d-4383-46d1-a0e0-6687f032c977 req-d4ab2b5c-3973-41cc-b682-ffca7c702429 service nova] Acquired lock "refresh_cache-b1d6409a-2733-470f-a929-672fe1631b1b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.534698] env[62692]: DEBUG nova.network.neutron [req-d982430d-4383-46d1-a0e0-6687f032c977 req-d4ab2b5c-3973-41cc-b682-ffca7c702429 service nova] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Refreshing network info cache for port ce7100ab-22d6-4268-8f4c-844d9658c6d1 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 826.582979] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 826.583407] env[62692]: DEBUG oslo_vmware.api [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52a38641-0524-66dc-d7cc-ade63a50539d, 'name': SearchDatastore_Task, 'duration_secs': 0.037954} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.583667] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 826.585292] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.585681] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] bb547773-d176-4c8e-a0fa-a374d5050b1b/81eb64fb-5a0c-47b7-9948-7dcb01b911bd-rescue.vmdk. {{(pid=62692) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 826.586059] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e8fd476b-7161-4b0a-8943-a5505ca351dc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.597161] env[62692]: DEBUG oslo_vmware.api [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 826.597161] env[62692]: value = "task-1141081" [ 826.597161] env[62692]: _type = "Task" [ 826.597161] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.610424] env[62692]: DEBUG oslo_vmware.api [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141081, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.636193] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141074, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.659545] env[62692]: DEBUG nova.scheduler.client.report [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 826.695340] env[62692]: DEBUG oslo_concurrency.lockutils [None req-73f17a3a-5c39-4408-9618-1d991f237d6b tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Lock "7a3f9c7d-03cf-4177-88db-acf33cc03feb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 119.218s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.752382] env[62692]: DEBUG oslo_vmware.api [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Task: {'id': task-1141077, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.173986} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.755011] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 826.755011] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dfb91df-6003-47f5-a9f8-7dc5b362234b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.783046] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Reconfiguring VM instance instance-00000030 to attach disk [datastore2] 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99/1cdd1dbc-567d-46f6-9f05-dfbc57acbf99.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 826.783934] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-15ff3de0-f22f-4cf0-b250-f327ffee4003 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.806708] env[62692]: DEBUG oslo_vmware.api [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Waiting for the task: (returnval){ [ 826.806708] env[62692]: value = "task-1141082" [ 826.806708] env[62692]: _type = "Task" [ 826.806708] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.815820] env[62692]: DEBUG oslo_vmware.api [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Task: {'id': task-1141082, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.867944] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-6240b800-2ca6-4456-90a7-7097659c4b52 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 826.868268] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dddf5ac9-4c3c-42e5-9c10-c663660eff05 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.879024] env[62692]: DEBUG oslo_vmware.api [None req-6240b800-2ca6-4456-90a7-7097659c4b52 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 826.879024] env[62692]: value = "task-1141083" [ 826.879024] env[62692]: _type = "Task" [ 826.879024] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.889888] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141080, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.893306] env[62692]: DEBUG oslo_vmware.api [None req-6240b800-2ca6-4456-90a7-7097659c4b52 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141083, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.993715] env[62692]: DEBUG nova.compute.manager [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 827.000033] env[62692]: DEBUG oslo_vmware.api [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141076, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.562775} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.000243] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 827.000432] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 827.000587] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 827.000947] env[62692]: INFO nova.compute.manager [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Took 1.81 seconds to destroy the instance on the hypervisor. [ 827.001128] env[62692]: DEBUG oslo.service.loopingcall [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 827.001327] env[62692]: DEBUG nova.compute.manager [-] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 827.001420] env[62692]: DEBUG nova.network.neutron [-] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 827.023557] env[62692]: DEBUG nova.virt.hardware [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 827.023927] env[62692]: DEBUG nova.virt.hardware [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 827.024199] env[62692]: DEBUG nova.virt.hardware [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 827.024479] env[62692]: DEBUG nova.virt.hardware [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 827.024705] env[62692]: DEBUG nova.virt.hardware [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 827.024934] env[62692]: DEBUG nova.virt.hardware [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 827.025336] env[62692]: DEBUG nova.virt.hardware [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 827.025507] env[62692]: DEBUG nova.virt.hardware [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 827.025765] env[62692]: DEBUG nova.virt.hardware [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 827.026032] env[62692]: DEBUG nova.virt.hardware [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 827.026308] env[62692]: DEBUG nova.virt.hardware [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 827.028368] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b64d71b5-c502-41c4-b8d8-6b4b5f0868d1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.043864] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed4f6529-fd74-4509-924b-05b82b8224ec {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.095114] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 827.095313] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Starting heal instance info cache {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 827.095444] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Rebuilding the list of instances to heal {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 827.108113] env[62692]: DEBUG oslo_vmware.api [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141081, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.137579] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141074, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.031865} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.137579] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 1848c251-0a0c-4c36-8bd0-ff37befde2a0/1848c251-0a0c-4c36-8bd0-ff37befde2a0.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 827.137579] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 827.137579] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ec41c737-bfee-484e-bae3-081f612f9a48 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.150212] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Waiting for the task: (returnval){ [ 827.150212] env[62692]: value = "task-1141084" [ 827.150212] env[62692]: _type = "Task" [ 827.150212] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.160450] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141084, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.167464] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.215s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.170119] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.816s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.170905] env[62692]: DEBUG nova.objects.instance [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lazy-loading 'resources' on Instance uuid fa7bb883-0065-4ecc-9eb5-8c5443344a2a {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 827.200980] env[62692]: INFO nova.scheduler.client.report [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Deleted allocations for instance 105e6fd5-4eff-4a0c-9a4b-e6deade781d5 [ 827.202618] env[62692]: DEBUG nova.compute.manager [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 827.322671] env[62692]: DEBUG oslo_vmware.api [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Task: {'id': task-1141082, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.387677] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141080, 'name': CreateVM_Task, 'duration_secs': 0.648593} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.388730] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 827.389680] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.390013] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.390370] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 827.390656] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-584c3249-432b-4d79-af92-067045e2f7af {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.399082] env[62692]: DEBUG oslo_vmware.api [None req-6240b800-2ca6-4456-90a7-7097659c4b52 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141083, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.403500] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 827.403500] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]523b6fb5-77f8-c2ed-0a5d-fd87b96f4081" [ 827.403500] env[62692]: _type = "Task" [ 827.403500] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.413957] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]523b6fb5-77f8-c2ed-0a5d-fd87b96f4081, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.579516] env[62692]: INFO nova.compute.manager [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Rebuilding instance [ 827.601012] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Skipping network cache update for instance because it is being deleted. {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9943}} [ 827.601671] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Skipping network cache update for instance because it is Building. {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 827.601992] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Skipping network cache update for instance because it is Building. {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 827.602271] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Skipping network cache update for instance because it is Building. {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 827.602620] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "refresh_cache-d99737d8-2eb0-40ee-b61e-6c736c84ea59" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.602887] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquired lock "refresh_cache-d99737d8-2eb0-40ee-b61e-6c736c84ea59" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.603156] env[62692]: DEBUG nova.network.neutron [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Forcefully refreshing network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 827.603357] env[62692]: DEBUG nova.objects.instance [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lazy-loading 'info_cache' on Instance uuid d99737d8-2eb0-40ee-b61e-6c736c84ea59 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 827.625703] env[62692]: DEBUG oslo_vmware.api [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141081, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.91086} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.626756] env[62692]: INFO nova.virt.vmwareapi.ds_util [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] bb547773-d176-4c8e-a0fa-a374d5050b1b/81eb64fb-5a0c-47b7-9948-7dcb01b911bd-rescue.vmdk. [ 827.631459] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-027b44f3-0175-4f9e-8c4d-bbcf47cfffcc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.660338] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Reconfiguring VM instance instance-00000029 to attach disk [datastore2] bb547773-d176-4c8e-a0fa-a374d5050b1b/81eb64fb-5a0c-47b7-9948-7dcb01b911bd-rescue.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 827.666333] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-81465baa-e607-4e06-87c0-98de47dca8c8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.684574] env[62692]: DEBUG nova.compute.manager [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 827.686556] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4b04c5c-0a0c-403f-bda8-e1e89c37cc92 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.692449] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141084, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.317293} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.694359] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 827.694793] env[62692]: DEBUG oslo_vmware.api [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 827.694793] env[62692]: value = "task-1141085" [ 827.694793] env[62692]: _type = "Task" [ 827.694793] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.695689] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eafeb95d-e0e4-46b7-ba54-2a3d40f5ba61 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.732370] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Reconfiguring VM instance instance-0000002e to attach disk [datastore1] 1848c251-0a0c-4c36-8bd0-ff37befde2a0/1848c251-0a0c-4c36-8bd0-ff37befde2a0.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 827.738822] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c4e9a1a1-608c-4f16-8f40-b55a4548d1d0 tempest-ServerDiagnosticsNegativeTest-1298436391 tempest-ServerDiagnosticsNegativeTest-1298436391-project-member] Lock "105e6fd5-4eff-4a0c-9a4b-e6deade781d5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.341s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.742009] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-25ee3c70-095a-4364-be96-e20b91199da3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.757448] env[62692]: DEBUG oslo_vmware.api [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141085, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.765335] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Waiting for the task: (returnval){ [ 827.765335] env[62692]: value = "task-1141086" [ 827.765335] env[62692]: _type = "Task" [ 827.765335] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.772102] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.776869] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141086, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.820267] env[62692]: DEBUG oslo_vmware.api [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Task: {'id': task-1141082, 'name': ReconfigVM_Task, 'duration_secs': 0.719665} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.820546] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Reconfigured VM instance instance-00000030 to attach disk [datastore2] 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99/1cdd1dbc-567d-46f6-9f05-dfbc57acbf99.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 827.821196] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c37ccbd8-de29-4e64-8870-1f7936d084eb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.823962] env[62692]: DEBUG nova.network.neutron [req-d982430d-4383-46d1-a0e0-6687f032c977 req-d4ab2b5c-3973-41cc-b682-ffca7c702429 service nova] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Updated VIF entry in instance network info cache for port ce7100ab-22d6-4268-8f4c-844d9658c6d1. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 827.824039] env[62692]: DEBUG nova.network.neutron [req-d982430d-4383-46d1-a0e0-6687f032c977 req-d4ab2b5c-3973-41cc-b682-ffca7c702429 service nova] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Updating instance_info_cache with network_info: [{"id": "ce7100ab-22d6-4268-8f4c-844d9658c6d1", "address": "fa:16:3e:59:21:56", "network": {"id": "dd260020-2d74-46ba-aa78-af29bdefa176", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-617179655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "589895ae20394e2fa525707bf7d09f2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce7100ab-22", "ovs_interfaceid": "ce7100ab-22d6-4268-8f4c-844d9658c6d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.828725] env[62692]: DEBUG oslo_vmware.api [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Waiting for the task: (returnval){ [ 827.828725] env[62692]: value = "task-1141087" [ 827.828725] env[62692]: _type = "Task" [ 827.828725] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.843451] env[62692]: DEBUG oslo_vmware.api [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Task: {'id': task-1141087, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.890406] env[62692]: DEBUG oslo_vmware.api [None req-6240b800-2ca6-4456-90a7-7097659c4b52 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141083, 'name': PowerOnVM_Task, 'duration_secs': 0.675432} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.892938] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-6240b800-2ca6-4456-90a7-7097659c4b52 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 827.892938] env[62692]: DEBUG nova.compute.manager [None req-6240b800-2ca6-4456-90a7-7097659c4b52 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 827.894340] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05971687-7b00-4319-8c9b-6691bb2b232b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.916828] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]523b6fb5-77f8-c2ed-0a5d-fd87b96f4081, 'name': SearchDatastore_Task, 'duration_secs': 0.05606} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.917804] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.918114] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 827.918350] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.919889] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.919889] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 827.919889] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2f725956-4c87-49ce-9174-b8963849ced1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.932435] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 827.932670] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 827.936863] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd136bfe-c727-4c67-8f51-67854fe0ae09 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.944686] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 827.944686] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52247776-9b90-9761-0038-6dae81414419" [ 827.944686] env[62692]: _type = "Task" [ 827.944686] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.956294] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52247776-9b90-9761-0038-6dae81414419, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.101721] env[62692]: DEBUG nova.network.neutron [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Successfully updated port: d996a5d5-fb43-4af0-9531-397082ddca3f {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 828.211944] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 828.212270] env[62692]: DEBUG oslo_vmware.api [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141085, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.214928] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-779a9c3f-3e43-4ab0-a6b5-4afe99cf4fa1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.224155] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Waiting for the task: (returnval){ [ 828.224155] env[62692]: value = "task-1141088" [ 828.224155] env[62692]: _type = "Task" [ 828.224155] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.242795] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141088, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.268583] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4006138a-6beb-4e3d-8d88-fb50b42c0067 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.282448] env[62692]: DEBUG nova.compute.manager [req-b8ff5a94-2ddc-45ac-bf4d-95840d0bd966 req-f685101d-05a2-446d-a07e-cbdc5a80dfe5 service nova] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Received event network-vif-plugged-d996a5d5-fb43-4af0-9531-397082ddca3f {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 828.282830] env[62692]: DEBUG oslo_concurrency.lockutils [req-b8ff5a94-2ddc-45ac-bf4d-95840d0bd966 req-f685101d-05a2-446d-a07e-cbdc5a80dfe5 service nova] Acquiring lock "ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.283188] env[62692]: DEBUG oslo_concurrency.lockutils [req-b8ff5a94-2ddc-45ac-bf4d-95840d0bd966 req-f685101d-05a2-446d-a07e-cbdc5a80dfe5 service nova] Lock "ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.283468] env[62692]: DEBUG oslo_concurrency.lockutils [req-b8ff5a94-2ddc-45ac-bf4d-95840d0bd966 req-f685101d-05a2-446d-a07e-cbdc5a80dfe5 service nova] Lock "ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.283726] env[62692]: DEBUG nova.compute.manager [req-b8ff5a94-2ddc-45ac-bf4d-95840d0bd966 req-f685101d-05a2-446d-a07e-cbdc5a80dfe5 service nova] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] No waiting events found dispatching network-vif-plugged-d996a5d5-fb43-4af0-9531-397082ddca3f {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 828.283992] env[62692]: WARNING nova.compute.manager [req-b8ff5a94-2ddc-45ac-bf4d-95840d0bd966 req-f685101d-05a2-446d-a07e-cbdc5a80dfe5 service nova] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Received unexpected event network-vif-plugged-d996a5d5-fb43-4af0-9531-397082ddca3f for instance with vm_state building and task_state spawning. [ 828.290307] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fa97655-88ae-4f39-92e0-4fcd04fb55d3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.307555] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141086, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.338797] env[62692]: DEBUG oslo_concurrency.lockutils [req-d982430d-4383-46d1-a0e0-6687f032c977 req-d4ab2b5c-3973-41cc-b682-ffca7c702429 service nova] Releasing lock "refresh_cache-b1d6409a-2733-470f-a929-672fe1631b1b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.344902] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a8974d2-d2d8-4728-91c9-2f3db73162dc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.356387] env[62692]: DEBUG oslo_vmware.api [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Task: {'id': task-1141087, 'name': Rename_Task, 'duration_secs': 0.334994} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.357107] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 828.358372] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cee5f152-a64a-45f4-a1b7-f1f6f3d43d22 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.362981] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2d85ef26-df0a-471c-a6e7-2ffc0a363984 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.377447] env[62692]: DEBUG nova.compute.provider_tree [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 828.380856] env[62692]: DEBUG oslo_vmware.api [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Waiting for the task: (returnval){ [ 828.380856] env[62692]: value = "task-1141089" [ 828.380856] env[62692]: _type = "Task" [ 828.380856] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.391385] env[62692]: DEBUG oslo_vmware.api [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Task: {'id': task-1141089, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.457253] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52247776-9b90-9761-0038-6dae81414419, 'name': SearchDatastore_Task, 'duration_secs': 0.025068} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.458192] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f77f4eb9-908b-4a88-897c-f8764429e6f3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.465359] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 828.465359] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]522b6871-7c51-5fef-3dd8-dee64977972f" [ 828.465359] env[62692]: _type = "Task" [ 828.465359] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.474484] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]522b6871-7c51-5fef-3dd8-dee64977972f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.603708] env[62692]: DEBUG oslo_concurrency.lockutils [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Acquiring lock "refresh_cache-ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.603914] env[62692]: DEBUG oslo_concurrency.lockutils [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Acquired lock "refresh_cache-ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.604189] env[62692]: DEBUG nova.network.neutron [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 828.633807] env[62692]: DEBUG nova.compute.manager [req-e229d19e-cd67-42d3-a66c-8ff5d46a8069 req-7bff337b-55e7-4fc7-9f3d-1aea83bb83f5 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Received event network-vif-deleted-b0db2db5-00e9-4a3a-a41a-7ed7f4985073 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 828.633996] env[62692]: INFO nova.compute.manager [req-e229d19e-cd67-42d3-a66c-8ff5d46a8069 req-7bff337b-55e7-4fc7-9f3d-1aea83bb83f5 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Neutron deleted interface b0db2db5-00e9-4a3a-a41a-7ed7f4985073; detaching it from the instance and deleting it from the info cache [ 828.634413] env[62692]: DEBUG nova.network.neutron [req-e229d19e-cd67-42d3-a66c-8ff5d46a8069 req-7bff337b-55e7-4fc7-9f3d-1aea83bb83f5 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Updating instance_info_cache with network_info: [{"id": "13349e97-1acd-48d2-a24b-44160a60773e", "address": "fa:16:3e:b7:2e:8e", "network": {"id": "6ad813f0-48dd-45af-89d3-62d3ee8d084f", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-738939750", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.66", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4fed583a514c4043a4b5772c73e21aff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "233536d0-6913-4879-8442-42dcf1d4ecbb", "external-id": "nsx-vlan-transportzone-700", "segmentation_id": 700, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13349e97-1a", "ovs_interfaceid": "13349e97-1acd-48d2-a24b-44160a60773e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "56f47df0-50f1-43e0-a2d5-3a4ad9d19327", "address": "fa:16:3e:35:ae:c9", "network": {"id": "5f63a59b-8a80-45bb-8b7f-23bfc180fe98", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1827914724", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.174", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fed583a514c4043a4b5772c73e21aff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d12aff80-9d1b-4a67-a470-9c0148b443e3", "external-id": "nsx-vlan-transportzone-784", "segmentation_id": 784, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56f47df0-50", "ovs_interfaceid": "56f47df0-50f1-43e0-a2d5-3a4ad9d19327", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.711374] env[62692]: DEBUG oslo_vmware.api [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141085, 'name': ReconfigVM_Task, 'duration_secs': 0.773232} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.711666] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Reconfigured VM instance instance-00000029 to attach disk [datastore2] bb547773-d176-4c8e-a0fa-a374d5050b1b/81eb64fb-5a0c-47b7-9948-7dcb01b911bd-rescue.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 828.712611] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a812af5-58e0-4c32-9d64-344f4f6068f6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.745772] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-776793d3-aa00-4aea-a4ea-5673d5933024 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.764045] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141088, 'name': PowerOffVM_Task, 'duration_secs': 0.244525} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.765885] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 828.765885] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 828.766178] env[62692]: DEBUG oslo_vmware.api [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 828.766178] env[62692]: value = "task-1141090" [ 828.766178] env[62692]: _type = "Task" [ 828.766178] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.766977] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63c2c7ff-f115-45c7-bef4-21d46dbfb774 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.782870] env[62692]: DEBUG oslo_vmware.api [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141090, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.787068] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 828.787365] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141086, 'name': ReconfigVM_Task, 'duration_secs': 0.885942} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.787571] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-689f239e-8b42-4b8f-90fa-bb78ae4502f2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.789305] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Reconfigured VM instance instance-0000002e to attach disk [datastore1] 1848c251-0a0c-4c36-8bd0-ff37befde2a0/1848c251-0a0c-4c36-8bd0-ff37befde2a0.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 828.789936] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9d4e594c-66b0-49d7-a577-d936ff1b776c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.802578] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Waiting for the task: (returnval){ [ 828.802578] env[62692]: value = "task-1141091" [ 828.802578] env[62692]: _type = "Task" [ 828.802578] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.814304] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141091, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.825892] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 828.825892] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 828.825892] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Deleting the datastore file [datastore1] 7a3f9c7d-03cf-4177-88db-acf33cc03feb {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 828.825892] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7ce8c20c-f39c-4b58-8e4e-e6081b9c29b3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.832464] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Waiting for the task: (returnval){ [ 828.832464] env[62692]: value = "task-1141093" [ 828.832464] env[62692]: _type = "Task" [ 828.832464] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.843182] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141093, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.848845] env[62692]: DEBUG nova.network.neutron [-] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.881109] env[62692]: DEBUG nova.scheduler.client.report [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 828.897767] env[62692]: DEBUG oslo_vmware.api [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Task: {'id': task-1141089, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.977840] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]522b6871-7c51-5fef-3dd8-dee64977972f, 'name': SearchDatastore_Task, 'duration_secs': 0.010765} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.977840] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.977840] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] b1d6409a-2733-470f-a929-672fe1631b1b/b1d6409a-2733-470f-a929-672fe1631b1b.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 828.978137] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e1fae283-c0ed-4458-8165-25134cf9534d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.985379] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 828.985379] env[62692]: value = "task-1141094" [ 828.985379] env[62692]: _type = "Task" [ 828.985379] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.994697] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141094, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.140549] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6d502444-28ae-4748-8beb-f77ee81b71ec {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.148830] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e61731cc-c078-4aa4-b39e-140be2c6ed88 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.166488] env[62692]: DEBUG nova.network.neutron [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 829.185334] env[62692]: DEBUG nova.compute.manager [req-e229d19e-cd67-42d3-a66c-8ff5d46a8069 req-7bff337b-55e7-4fc7-9f3d-1aea83bb83f5 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Detach interface failed, port_id=b0db2db5-00e9-4a3a-a41a-7ed7f4985073, reason: Instance aa0666ce-4d4f-4333-8acb-ea0cfb438cd9 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 829.185588] env[62692]: DEBUG nova.compute.manager [req-e229d19e-cd67-42d3-a66c-8ff5d46a8069 req-7bff337b-55e7-4fc7-9f3d-1aea83bb83f5 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Received event network-vif-deleted-56f47df0-50f1-43e0-a2d5-3a4ad9d19327 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 829.185768] env[62692]: INFO nova.compute.manager [req-e229d19e-cd67-42d3-a66c-8ff5d46a8069 req-7bff337b-55e7-4fc7-9f3d-1aea83bb83f5 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Neutron deleted interface 56f47df0-50f1-43e0-a2d5-3a4ad9d19327; detaching it from the instance and deleting it from the info cache [ 829.186033] env[62692]: DEBUG nova.network.neutron [req-e229d19e-cd67-42d3-a66c-8ff5d46a8069 req-7bff337b-55e7-4fc7-9f3d-1aea83bb83f5 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Updating instance_info_cache with network_info: [{"id": "13349e97-1acd-48d2-a24b-44160a60773e", "address": "fa:16:3e:b7:2e:8e", "network": {"id": "6ad813f0-48dd-45af-89d3-62d3ee8d084f", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-738939750", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.66", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4fed583a514c4043a4b5772c73e21aff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "233536d0-6913-4879-8442-42dcf1d4ecbb", "external-id": "nsx-vlan-transportzone-700", "segmentation_id": 700, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13349e97-1a", "ovs_interfaceid": "13349e97-1acd-48d2-a24b-44160a60773e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.285189] env[62692]: DEBUG oslo_vmware.api [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141090, 'name': ReconfigVM_Task, 'duration_secs': 0.172476} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.285189] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 829.285189] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-71c9585c-44bd-4567-a990-78eb5f975fb7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.295420] env[62692]: DEBUG oslo_vmware.api [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 829.295420] env[62692]: value = "task-1141095" [ 829.295420] env[62692]: _type = "Task" [ 829.295420] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.310749] env[62692]: DEBUG oslo_vmware.api [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141095, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.316874] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141091, 'name': Rename_Task, 'duration_secs': 0.332478} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.319526] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 829.319821] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ad96047b-f59b-440a-904d-8be389998930 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.334455] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Waiting for the task: (returnval){ [ 829.334455] env[62692]: value = "task-1141096" [ 829.334455] env[62692]: _type = "Task" [ 829.334455] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.351855] env[62692]: INFO nova.compute.manager [-] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Took 2.35 seconds to deallocate network for instance. [ 829.352454] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141093, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.125292} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.359172] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 829.359504] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 829.359581] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 829.361996] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141096, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.378082] env[62692]: DEBUG nova.network.neutron [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Updating instance_info_cache with network_info: [{"id": "d996a5d5-fb43-4af0-9531-397082ddca3f", "address": "fa:16:3e:a1:e9:75", "network": {"id": "19d432d2-7d87-4b3d-a365-9ca0ebb38db5", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1014257201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "454fce691f174c7e9348638d98597a26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48937bd1-23dc-413f-b46b-59bf9e709aa0", "external-id": "nsx-vlan-transportzone-160", "segmentation_id": 160, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd996a5d5-fb", "ovs_interfaceid": "d996a5d5-fb43-4af0-9531-397082ddca3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.386935] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.217s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.390429] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.845s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.392201] env[62692]: INFO nova.compute.claims [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 829.407097] env[62692]: DEBUG oslo_vmware.api [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Task: {'id': task-1141089, 'name': PowerOnVM_Task, 'duration_secs': 0.911498} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.408459] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 829.408580] env[62692]: INFO nova.compute.manager [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Took 10.37 seconds to spawn the instance on the hypervisor. [ 829.408968] env[62692]: DEBUG nova.compute.manager [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 829.409845] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36ceca37-3d12-4775-b568-088068d912c8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.425899] env[62692]: INFO nova.scheduler.client.report [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Deleted allocations for instance fa7bb883-0065-4ecc-9eb5-8c5443344a2a [ 829.489862] env[62692]: DEBUG nova.network.neutron [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Updating instance_info_cache with network_info: [{"id": "7f2cabdf-cbda-4cf7-a518-13f619716d12", "address": "fa:16:3e:6b:22:0c", "network": {"id": "f70727d6-a9aa-44b0-a244-2533b3f96571", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-284105837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1037e5d436bd429391159b87a712f6e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f2cabdf-cb", "ovs_interfaceid": "7f2cabdf-cbda-4cf7-a518-13f619716d12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.503025] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141094, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.691411] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dd693e66-c3d3-41d6-95a0-c0e062c159b9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.699581] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ea5a4de-7d40-4b9f-9419-755e85866fa7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.738202] env[62692]: DEBUG nova.compute.manager [req-e229d19e-cd67-42d3-a66c-8ff5d46a8069 req-7bff337b-55e7-4fc7-9f3d-1aea83bb83f5 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Detach interface failed, port_id=56f47df0-50f1-43e0-a2d5-3a4ad9d19327, reason: Instance aa0666ce-4d4f-4333-8acb-ea0cfb438cd9 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 829.738561] env[62692]: DEBUG nova.compute.manager [req-e229d19e-cd67-42d3-a66c-8ff5d46a8069 req-7bff337b-55e7-4fc7-9f3d-1aea83bb83f5 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Received event network-vif-deleted-13349e97-1acd-48d2-a24b-44160a60773e {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 829.738837] env[62692]: INFO nova.compute.manager [req-e229d19e-cd67-42d3-a66c-8ff5d46a8069 req-7bff337b-55e7-4fc7-9f3d-1aea83bb83f5 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Neutron deleted interface 13349e97-1acd-48d2-a24b-44160a60773e; detaching it from the instance and deleting it from the info cache [ 829.739091] env[62692]: DEBUG nova.network.neutron [req-e229d19e-cd67-42d3-a66c-8ff5d46a8069 req-7bff337b-55e7-4fc7-9f3d-1aea83bb83f5 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.806281] env[62692]: DEBUG oslo_vmware.api [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141095, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.851920] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141096, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.869069] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.880882] env[62692]: DEBUG oslo_concurrency.lockutils [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Releasing lock "refresh_cache-ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.881428] env[62692]: DEBUG nova.compute.manager [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Instance network_info: |[{"id": "d996a5d5-fb43-4af0-9531-397082ddca3f", "address": "fa:16:3e:a1:e9:75", "network": {"id": "19d432d2-7d87-4b3d-a365-9ca0ebb38db5", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1014257201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "454fce691f174c7e9348638d98597a26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48937bd1-23dc-413f-b46b-59bf9e709aa0", "external-id": "nsx-vlan-transportzone-160", "segmentation_id": 160, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd996a5d5-fb", "ovs_interfaceid": "d996a5d5-fb43-4af0-9531-397082ddca3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 829.881834] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a1:e9:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '48937bd1-23dc-413f-b46b-59bf9e709aa0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd996a5d5-fb43-4af0-9531-397082ddca3f', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 829.889455] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Creating folder: Project (454fce691f174c7e9348638d98597a26). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 829.890157] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f9f0ac33-38d3-471f-8cf9-e7d612055ead {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.904693] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Created folder: Project (454fce691f174c7e9348638d98597a26) in parent group-v248868. [ 829.905263] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Creating folder: Instances. Parent ref: group-v248931. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 829.905374] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-416e3305-feea-4e2c-9562-35000d277776 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.917977] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Created folder: Instances in parent group-v248931. [ 829.918342] env[62692]: DEBUG oslo.service.loopingcall [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 829.918604] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 829.918896] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6dea9332-24c8-449c-bef5-3b19577c578c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.942209] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5c73f2ee-2b5e-47fc-b26a-3320d8f1b8aa tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "fa7bb883-0065-4ecc-9eb5-8c5443344a2a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.568s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.943278] env[62692]: INFO nova.compute.manager [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Took 28.84 seconds to build instance. [ 829.948049] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 829.948049] env[62692]: value = "task-1141099" [ 829.948049] env[62692]: _type = "Task" [ 829.948049] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.957736] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141099, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.996239] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Releasing lock "refresh_cache-d99737d8-2eb0-40ee-b61e-6c736c84ea59" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.996452] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Updated the network info_cache for instance {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 829.996696] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 829.997434] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 829.997640] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 829.997799] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 829.998089] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 829.998399] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 829.998589] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62692) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 829.998883] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager.update_available_resource {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 830.005683] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141094, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.243421] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3d41de42-6697-4512-8342-29117d181ca6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.261828] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e7e3115-ad7d-46a2-ab0b-4c786191ebc8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.312739] env[62692]: DEBUG nova.compute.manager [req-e229d19e-cd67-42d3-a66c-8ff5d46a8069 req-7bff337b-55e7-4fc7-9f3d-1aea83bb83f5 service nova] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Detach interface failed, port_id=13349e97-1acd-48d2-a24b-44160a60773e, reason: Instance aa0666ce-4d4f-4333-8acb-ea0cfb438cd9 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 830.332032] env[62692]: DEBUG oslo_vmware.api [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141095, 'name': PowerOnVM_Task} progress is 71%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.356389] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141096, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.371145] env[62692]: DEBUG nova.compute.manager [req-47c03c2d-79e4-4fbd-9d1b-47ab4671522b req-eb4afefa-e560-48b0-b623-c5700b56f362 service nova] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Received event network-changed-d996a5d5-fb43-4af0-9531-397082ddca3f {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 830.371398] env[62692]: DEBUG nova.compute.manager [req-47c03c2d-79e4-4fbd-9d1b-47ab4671522b req-eb4afefa-e560-48b0-b623-c5700b56f362 service nova] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Refreshing instance network info cache due to event network-changed-d996a5d5-fb43-4af0-9531-397082ddca3f. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 830.371638] env[62692]: DEBUG oslo_concurrency.lockutils [req-47c03c2d-79e4-4fbd-9d1b-47ab4671522b req-eb4afefa-e560-48b0-b623-c5700b56f362 service nova] Acquiring lock "refresh_cache-ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.371777] env[62692]: DEBUG oslo_concurrency.lockutils [req-47c03c2d-79e4-4fbd-9d1b-47ab4671522b req-eb4afefa-e560-48b0-b623-c5700b56f362 service nova] Acquired lock "refresh_cache-ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.371954] env[62692]: DEBUG nova.network.neutron [req-47c03c2d-79e4-4fbd-9d1b-47ab4671522b req-eb4afefa-e560-48b0-b623-c5700b56f362 service nova] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Refreshing network info cache for port d996a5d5-fb43-4af0-9531-397082ddca3f {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 830.411701] env[62692]: DEBUG nova.virt.hardware [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 830.412017] env[62692]: DEBUG nova.virt.hardware [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 830.412220] env[62692]: DEBUG nova.virt.hardware [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 830.412462] env[62692]: DEBUG nova.virt.hardware [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 830.412645] env[62692]: DEBUG nova.virt.hardware [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 830.412825] env[62692]: DEBUG nova.virt.hardware [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 830.416146] env[62692]: DEBUG nova.virt.hardware [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 830.416399] env[62692]: DEBUG nova.virt.hardware [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 830.416642] env[62692]: DEBUG nova.virt.hardware [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 830.416867] env[62692]: DEBUG nova.virt.hardware [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 830.417073] env[62692]: DEBUG nova.virt.hardware [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 830.418475] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-736953ee-4e5d-4a04-87df-c594a88a9b42 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.427860] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-933076d2-e694-4668-a61a-a73be77a6482 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.445458] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e7096c3-1d4e-469b-adce-704282ad09c0 tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Lock "1cdd1dbc-567d-46f6-9f05-dfbc57acbf99" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.020s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.446198] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Instance VIF info [] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 830.451875] env[62692]: DEBUG oslo.service.loopingcall [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.456040] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 830.460063] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ea103022-aa51-47b4-871e-da67312e532d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.483205] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141099, 'name': CreateVM_Task, 'duration_secs': 0.464414} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.484499] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 830.484722] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 830.484722] env[62692]: value = "task-1141100" [ 830.484722] env[62692]: _type = "Task" [ 830.484722] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.485434] env[62692]: DEBUG oslo_concurrency.lockutils [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.485770] env[62692]: DEBUG oslo_concurrency.lockutils [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.485917] env[62692]: DEBUG oslo_concurrency.lockutils [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 830.488775] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85266384-b947-4b65-bf1b-462a12c40dfb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.505985] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Waiting for the task: (returnval){ [ 830.505985] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52579ba7-9448-3845-a6aa-146da08eb0df" [ 830.505985] env[62692]: _type = "Task" [ 830.505985] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.506277] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141100, 'name': CreateVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.510637] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.519619] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141094, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.421141} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.520352] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] b1d6409a-2733-470f-a929-672fe1631b1b/b1d6409a-2733-470f-a929-672fe1631b1b.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 830.520582] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 830.520847] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4cb7fe12-94a7-4bc3-a95a-4979bf92b467 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.529411] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52579ba7-9448-3845-a6aa-146da08eb0df, 'name': SearchDatastore_Task, 'duration_secs': 0.013181} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.529982] env[62692]: DEBUG oslo_concurrency.lockutils [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.530243] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 830.530474] env[62692]: DEBUG oslo_concurrency.lockutils [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.530661] env[62692]: DEBUG oslo_concurrency.lockutils [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.530919] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 830.532522] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1287c82e-240f-4871-a656-22fa28b0e080 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.535669] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 830.535669] env[62692]: value = "task-1141101" [ 830.535669] env[62692]: _type = "Task" [ 830.535669] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.544962] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 830.545322] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 830.552441] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aaedfea0-78e6-4d32-bf42-6979b2a6ca5a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.555214] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141101, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.559604] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Waiting for the task: (returnval){ [ 830.559604] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]521b7615-5b69-7fbc-bfc4-f4a083acfb61" [ 830.559604] env[62692]: _type = "Task" [ 830.559604] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.569832] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]521b7615-5b69-7fbc-bfc4-f4a083acfb61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.822030] env[62692]: DEBUG oslo_vmware.api [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141095, 'name': PowerOnVM_Task, 'duration_secs': 1.297733} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.822292] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 830.825897] env[62692]: DEBUG nova.compute.manager [None req-ce616ec2-8663-41de-b51d-10eb4368275d tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 830.826611] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e940e3d4-7322-4a0f-9810-7e229f226147 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.857515] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141096, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.902929] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbbcb927-0533-4768-8f57-552c57688a12 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.911894] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6c7cb3e-0793-4750-8781-d9abab2c2bc9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.943854] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ec58a4f-80a2-49cb-8865-13d5fe5622bc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.952704] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-106dc0c3-fd6e-4405-b62a-d87dd66c52ce {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.972193] env[62692]: DEBUG nova.compute.provider_tree [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 830.977461] env[62692]: DEBUG nova.compute.manager [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 830.998306] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141100, 'name': CreateVM_Task, 'duration_secs': 0.455034} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.998478] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 830.998907] env[62692]: DEBUG oslo_concurrency.lockutils [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.999079] env[62692]: DEBUG oslo_concurrency.lockutils [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.999398] env[62692]: DEBUG oslo_concurrency.lockutils [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 830.999648] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79da7087-e6df-4538-840b-e92d51818d62 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.004954] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Waiting for the task: (returnval){ [ 831.004954] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]521c6e4d-2318-1b73-b163-8d2cc908d399" [ 831.004954] env[62692]: _type = "Task" [ 831.004954] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.010268] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Acquiring lock "1cdd1dbc-567d-46f6-9f05-dfbc57acbf99" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.010496] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Lock "1cdd1dbc-567d-46f6-9f05-dfbc57acbf99" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.010708] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Acquiring lock "1cdd1dbc-567d-46f6-9f05-dfbc57acbf99-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.010876] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Lock "1cdd1dbc-567d-46f6-9f05-dfbc57acbf99-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.011045] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Lock "1cdd1dbc-567d-46f6-9f05-dfbc57acbf99-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.015737] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]521c6e4d-2318-1b73-b163-8d2cc908d399, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.016218] env[62692]: INFO nova.compute.manager [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Terminating instance [ 831.018596] env[62692]: DEBUG nova.compute.manager [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 831.018734] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 831.019455] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48e26faa-5486-4f18-a6fb-96cd2a51fcd0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.027775] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 831.028044] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-23b6af62-7ce4-44b0-89c4-d2a159e7b943 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.035418] env[62692]: DEBUG oslo_vmware.api [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Waiting for the task: (returnval){ [ 831.035418] env[62692]: value = "task-1141102" [ 831.035418] env[62692]: _type = "Task" [ 831.035418] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.049731] env[62692]: DEBUG oslo_vmware.api [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Task: {'id': task-1141102, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.052826] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141101, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088294} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.053090] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 831.053884] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5c499bc-f101-4f02-b61a-4fba40ee349f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.077470] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] b1d6409a-2733-470f-a929-672fe1631b1b/b1d6409a-2733-470f-a929-672fe1631b1b.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 831.082705] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-42aa0e1b-f2af-4a0b-972a-810a1c6ca3fe {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.107492] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]521b7615-5b69-7fbc-bfc4-f4a083acfb61, 'name': SearchDatastore_Task, 'duration_secs': 0.012236} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.107492] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 831.107492] env[62692]: value = "task-1141103" [ 831.107492] env[62692]: _type = "Task" [ 831.107492] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.107492] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f886a82-e5dc-44fc-a986-4f54e86cad15 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.120697] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Waiting for the task: (returnval){ [ 831.120697] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52b030f5-aee5-c76b-dc50-50b7c944da49" [ 831.120697] env[62692]: _type = "Task" [ 831.120697] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.120859] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141103, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.130296] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52b030f5-aee5-c76b-dc50-50b7c944da49, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.247426] env[62692]: DEBUG nova.network.neutron [req-47c03c2d-79e4-4fbd-9d1b-47ab4671522b req-eb4afefa-e560-48b0-b623-c5700b56f362 service nova] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Updated VIF entry in instance network info cache for port d996a5d5-fb43-4af0-9531-397082ddca3f. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 831.247863] env[62692]: DEBUG nova.network.neutron [req-47c03c2d-79e4-4fbd-9d1b-47ab4671522b req-eb4afefa-e560-48b0-b623-c5700b56f362 service nova] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Updating instance_info_cache with network_info: [{"id": "d996a5d5-fb43-4af0-9531-397082ddca3f", "address": "fa:16:3e:a1:e9:75", "network": {"id": "19d432d2-7d87-4b3d-a365-9ca0ebb38db5", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1014257201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "454fce691f174c7e9348638d98597a26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48937bd1-23dc-413f-b46b-59bf9e709aa0", "external-id": "nsx-vlan-transportzone-160", "segmentation_id": 160, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd996a5d5-fb", "ovs_interfaceid": "d996a5d5-fb43-4af0-9531-397082ddca3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.353758] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141096, 'name': PowerOnVM_Task} progress is 82%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.481326] env[62692]: DEBUG nova.scheduler.client.report [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 831.512277] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.518613] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]521c6e4d-2318-1b73-b163-8d2cc908d399, 'name': SearchDatastore_Task, 'duration_secs': 0.046962} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.518938] env[62692]: DEBUG oslo_concurrency.lockutils [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.519197] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 831.519407] env[62692]: DEBUG oslo_concurrency.lockutils [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.552167] env[62692]: DEBUG oslo_vmware.api [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Task: {'id': task-1141102, 'name': PowerOffVM_Task, 'duration_secs': 0.48313} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.552167] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 831.552167] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 831.552167] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e09cedef-3cec-48e2-8b24-f9fe5bfd8068 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.623599] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 831.623599] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 831.623599] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Deleting the datastore file [datastore2] 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 831.626312] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d5146b4d-52b0-446a-89f3-9216296e2502 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.408795] env[62692]: DEBUG oslo_concurrency.lockutils [req-47c03c2d-79e4-4fbd-9d1b-47ab4671522b req-eb4afefa-e560-48b0-b623-c5700b56f362 service nova] Releasing lock "refresh_cache-ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.409503] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.019s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.409784] env[62692]: DEBUG nova.compute.manager [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 832.413122] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141103, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.413122] env[62692]: WARNING oslo_vmware.common.loopingcall [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] task run outlasted interval by 0.30098199999999997 sec [ 832.417857] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.563s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.421024] env[62692]: INFO nova.compute.claims [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 832.432637] env[62692]: DEBUG oslo_vmware.api [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Waiting for the task: (returnval){ [ 832.432637] env[62692]: value = "task-1141105" [ 832.432637] env[62692]: _type = "Task" [ 832.432637] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.432903] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52b030f5-aee5-c76b-dc50-50b7c944da49, 'name': SearchDatastore_Task, 'duration_secs': 0.036689} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.434135] env[62692]: DEBUG oslo_concurrency.lockutils [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.434373] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba/ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 832.438674] env[62692]: DEBUG oslo_concurrency.lockutils [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.438674] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 832.438674] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2ced59bc-c845-4fcc-8276-c0be3c721617 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.447041] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e74a514e-7d1b-4f1c-b99d-44ea8ffada5d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.448909] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141103, 'name': ReconfigVM_Task, 'duration_secs': 0.641472} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.449147] env[62692]: DEBUG oslo_vmware.api [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141096, 'name': PowerOnVM_Task, 'duration_secs': 2.208429} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.451385] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Reconfigured VM instance instance-00000032 to attach disk [datastore2] b1d6409a-2733-470f-a929-672fe1631b1b/b1d6409a-2733-470f-a929-672fe1631b1b.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 832.451803] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 832.452031] env[62692]: DEBUG nova.compute.manager [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 832.453213] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7e9a2998-a966-4d6e-88df-2673e3fa200f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.455345] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55fa32a5-4de3-436b-8893-2a82b8b45ef8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.463770] env[62692]: DEBUG oslo_vmware.api [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Task: {'id': task-1141105, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.467218] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 832.467396] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 832.468785] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cfdb2d2f-7530-4479-beda-f63b9a3d5988 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.474106] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 832.474106] env[62692]: value = "task-1141106" [ 832.474106] env[62692]: _type = "Task" [ 832.474106] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.476886] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Waiting for the task: (returnval){ [ 832.476886] env[62692]: value = "task-1141107" [ 832.476886] env[62692]: _type = "Task" [ 832.476886] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.491656] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Waiting for the task: (returnval){ [ 832.491656] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52fad93b-ce42-fff6-089b-711395c467fa" [ 832.491656] env[62692]: _type = "Task" [ 832.491656] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.502236] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Task: {'id': task-1141107, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.502475] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141106, 'name': Rename_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.509489] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52fad93b-ce42-fff6-089b-711395c467fa, 'name': SearchDatastore_Task, 'duration_secs': 0.010919} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.510662] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc21649a-6201-473f-aa04-1f5e6c686483 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.517270] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Waiting for the task: (returnval){ [ 832.517270] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]523a6155-3d40-074d-b747-191820c5dcf1" [ 832.517270] env[62692]: _type = "Task" [ 832.517270] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.527383] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]523a6155-3d40-074d-b747-191820c5dcf1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.931255] env[62692]: DEBUG nova.compute.utils [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 832.933184] env[62692]: DEBUG nova.compute.manager [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 832.933416] env[62692]: DEBUG nova.network.neutron [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 832.936606] env[62692]: INFO nova.compute.manager [None req-ff824c7d-c377-450a-be34-ad90487585f9 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Unrescuing [ 832.941355] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ff824c7d-c377-450a-be34-ad90487585f9 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "refresh_cache-bb547773-d176-4c8e-a0fa-a374d5050b1b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.941355] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ff824c7d-c377-450a-be34-ad90487585f9 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquired lock "refresh_cache-bb547773-d176-4c8e-a0fa-a374d5050b1b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.941355] env[62692]: DEBUG nova.network.neutron [None req-ff824c7d-c377-450a-be34-ad90487585f9 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 832.962350] env[62692]: DEBUG oslo_vmware.api [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Task: {'id': task-1141105, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.224778} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.962644] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 832.963264] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 832.963517] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 832.963764] env[62692]: INFO nova.compute.manager [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Took 1.95 seconds to destroy the instance on the hypervisor. [ 832.964024] env[62692]: DEBUG oslo.service.loopingcall [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 832.964236] env[62692]: DEBUG nova.compute.manager [-] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 832.964332] env[62692]: DEBUG nova.network.neutron [-] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 832.990347] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141106, 'name': Rename_Task, 'duration_secs': 0.1964} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.991977] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 832.997205] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d498070e-afed-46c4-b761-5276d0429226 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.012970] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Task: {'id': task-1141107, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.021132] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.021543] env[62692]: DEBUG nova.policy [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f2e2ee26163841ddb15bb67520e98394', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e13f8888d5e84eed8247101558e29a56', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 833.029055] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 833.029055] env[62692]: value = "task-1141108" [ 833.029055] env[62692]: _type = "Task" [ 833.029055] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.036586] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]523a6155-3d40-074d-b747-191820c5dcf1, 'name': SearchDatastore_Task, 'duration_secs': 0.016694} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.037617] env[62692]: DEBUG oslo_concurrency.lockutils [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.037893] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 7a3f9c7d-03cf-4177-88db-acf33cc03feb/7a3f9c7d-03cf-4177-88db-acf33cc03feb.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 833.038574] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9fa16b02-a899-465d-a35b-af8301d95a38 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.044733] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141108, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.046688] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Waiting for the task: (returnval){ [ 833.046688] env[62692]: value = "task-1141109" [ 833.046688] env[62692]: _type = "Task" [ 833.046688] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.056271] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141109, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.305034] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Acquiring lock "b389b14d-6dcc-4b26-84f8-38b952bf6ecd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.305034] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Lock "b389b14d-6dcc-4b26-84f8-38b952bf6ecd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.397379] env[62692]: DEBUG nova.compute.manager [req-552eab61-e30c-416e-ba32-a4f0e5689dbd req-efa27174-232a-4b72-9ecd-b2c8985f7aa9 service nova] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Received event network-vif-deleted-60ca6789-edcf-48cb-855a-52e119e461e5 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 833.397537] env[62692]: INFO nova.compute.manager [req-552eab61-e30c-416e-ba32-a4f0e5689dbd req-efa27174-232a-4b72-9ecd-b2c8985f7aa9 service nova] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Neutron deleted interface 60ca6789-edcf-48cb-855a-52e119e461e5; detaching it from the instance and deleting it from the info cache [ 833.397715] env[62692]: DEBUG nova.network.neutron [req-552eab61-e30c-416e-ba32-a4f0e5689dbd req-efa27174-232a-4b72-9ecd-b2c8985f7aa9 service nova] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.443934] env[62692]: DEBUG nova.compute.manager [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 833.466800] env[62692]: DEBUG nova.network.neutron [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Successfully created port: 4eb54953-e726-4bd7-9793-30a73f994d30 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 833.504568] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Task: {'id': task-1141107, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.60179} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.506557] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba/ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 833.506557] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 833.506926] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f4ffa949-2909-4144-826b-3f568e490734 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.523479] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Waiting for the task: (returnval){ [ 833.523479] env[62692]: value = "task-1141110" [ 833.523479] env[62692]: _type = "Task" [ 833.523479] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.547189] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141108, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.552075] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Task: {'id': task-1141110, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.567085] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141109, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.722164] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Acquiring lock "1848c251-0a0c-4c36-8bd0-ff37befde2a0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.722164] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Lock "1848c251-0a0c-4c36-8bd0-ff37befde2a0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.722164] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Acquiring lock "1848c251-0a0c-4c36-8bd0-ff37befde2a0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.722164] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Lock "1848c251-0a0c-4c36-8bd0-ff37befde2a0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.722164] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Lock "1848c251-0a0c-4c36-8bd0-ff37befde2a0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.730321] env[62692]: INFO nova.compute.manager [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Terminating instance [ 833.730937] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Acquiring lock "refresh_cache-1848c251-0a0c-4c36-8bd0-ff37befde2a0" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.731254] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Acquired lock "refresh_cache-1848c251-0a0c-4c36-8bd0-ff37befde2a0" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.731558] env[62692]: DEBUG nova.network.neutron [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 833.767270] env[62692]: DEBUG nova.network.neutron [-] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.861825] env[62692]: DEBUG nova.network.neutron [None req-ff824c7d-c377-450a-be34-ad90487585f9 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Updating instance_info_cache with network_info: [{"id": "c77af962-553d-4504-825e-78a6689aee79", "address": "fa:16:3e:9a:01:cc", "network": {"id": "17e72c3b-603f-4305-96ac-89a9b4c9c7e9", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-123501993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7db5465cb5db465b92dfa570572b6e80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc77af962-55", "ovs_interfaceid": "c77af962-553d-4504-825e-78a6689aee79", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.903951] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8977fd2b-372a-4b7a-9239-0087cb044bd8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.914307] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8dac476-4b5d-43c2-8d4e-652708471594 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.956981] env[62692]: DEBUG nova.compute.manager [req-552eab61-e30c-416e-ba32-a4f0e5689dbd req-efa27174-232a-4b72-9ecd-b2c8985f7aa9 service nova] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Detach interface failed, port_id=60ca6789-edcf-48cb-855a-52e119e461e5, reason: Instance 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 834.006647] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-847974e1-8916-413c-bb42-6e790799731a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.016867] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aaf6487-dae7-4b39-9191-1a04f5b2fb77 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.060018] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b824205-fe43-4af6-a884-b7d35f3aa8c5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.065272] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Task: {'id': task-1141110, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.11874} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.066524] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 834.067907] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09f3f289-d4d2-415c-9025-f87d04516db1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.077452] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141109, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.607936} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.077682] env[62692]: DEBUG oslo_vmware.api [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141108, 'name': PowerOnVM_Task, 'duration_secs': 0.536027} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.078364] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 7a3f9c7d-03cf-4177-88db-acf33cc03feb/7a3f9c7d-03cf-4177-88db-acf33cc03feb.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 834.078583] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 834.078870] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 834.079090] env[62692]: INFO nova.compute.manager [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Took 10.81 seconds to spawn the instance on the hypervisor. [ 834.079271] env[62692]: DEBUG nova.compute.manager [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 834.081224] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4137f62-d9d3-4439-a875-b2c4d2f11284 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.093342] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d09078e8-baf4-4fc5-bf7c-88001ff3dcaf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.095957] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a92f91e-eaf5-43bf-8401-96a64f43f7c5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.106814] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba/ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 834.107554] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-abe9f6d8-c865-4985-a588-3b230d840487 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.130956] env[62692]: DEBUG nova.compute.provider_tree [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 834.139581] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Waiting for the task: (returnval){ [ 834.139581] env[62692]: value = "task-1141111" [ 834.139581] env[62692]: _type = "Task" [ 834.139581] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.139953] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Waiting for the task: (returnval){ [ 834.139953] env[62692]: value = "task-1141112" [ 834.139953] env[62692]: _type = "Task" [ 834.139953] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.152936] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141111, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.156668] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Task: {'id': task-1141112, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.253075] env[62692]: DEBUG nova.network.neutron [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 834.270597] env[62692]: INFO nova.compute.manager [-] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Took 1.31 seconds to deallocate network for instance. [ 834.311061] env[62692]: DEBUG nova.network.neutron [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.366124] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ff824c7d-c377-450a-be34-ad90487585f9 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Releasing lock "refresh_cache-bb547773-d176-4c8e-a0fa-a374d5050b1b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.366928] env[62692]: DEBUG nova.objects.instance [None req-ff824c7d-c377-450a-be34-ad90487585f9 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lazy-loading 'flavor' on Instance uuid bb547773-d176-4c8e-a0fa-a374d5050b1b {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 834.463862] env[62692]: DEBUG nova.compute.manager [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 834.495409] env[62692]: DEBUG nova.virt.hardware [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 834.495683] env[62692]: DEBUG nova.virt.hardware [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 834.495846] env[62692]: DEBUG nova.virt.hardware [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 834.501020] env[62692]: DEBUG nova.virt.hardware [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 834.501020] env[62692]: DEBUG nova.virt.hardware [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 834.501020] env[62692]: DEBUG nova.virt.hardware [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 834.501020] env[62692]: DEBUG nova.virt.hardware [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 834.501020] env[62692]: DEBUG nova.virt.hardware [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 834.501020] env[62692]: DEBUG nova.virt.hardware [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 834.501020] env[62692]: DEBUG nova.virt.hardware [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 834.501020] env[62692]: DEBUG nova.virt.hardware [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 834.501020] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f3d6a75-a8b0-4b18-a88b-aeae124d74f0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.509648] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda6bdbe-b923-4627-b0fb-4742377fc825 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.658185] env[62692]: INFO nova.compute.manager [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Took 27.24 seconds to build instance. [ 834.673381] env[62692]: ERROR nova.scheduler.client.report [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [req-942dce43-1f6d-4694-b7fa-67cd51646169] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 470758a8-5efa-4875-aad5-f512a727752f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-942dce43-1f6d-4694-b7fa-67cd51646169"}]} [ 834.673817] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141111, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.162875} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.674048] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Task: {'id': task-1141112, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.677996] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 834.679272] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01bff382-4951-4f75-9530-8d9ad809c102 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.704114] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Reconfiguring VM instance instance-00000031 to attach disk [datastore2] 7a3f9c7d-03cf-4177-88db-acf33cc03feb/7a3f9c7d-03cf-4177-88db-acf33cc03feb.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 834.706276] env[62692]: DEBUG nova.scheduler.client.report [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Refreshing inventories for resource provider 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 834.709137] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-94be6074-fcd1-45ff-b8a6-0c877d9d477f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.732032] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Waiting for the task: (returnval){ [ 834.732032] env[62692]: value = "task-1141113" [ 834.732032] env[62692]: _type = "Task" [ 834.732032] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.740372] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141113, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.741512] env[62692]: DEBUG nova.scheduler.client.report [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Updating ProviderTree inventory for provider 470758a8-5efa-4875-aad5-f512a727752f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 834.741770] env[62692]: DEBUG nova.compute.provider_tree [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 834.761379] env[62692]: DEBUG nova.scheduler.client.report [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Refreshing aggregate associations for resource provider 470758a8-5efa-4875-aad5-f512a727752f, aggregates: None {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 834.777840] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.782939] env[62692]: DEBUG nova.scheduler.client.report [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Refreshing trait associations for resource provider 470758a8-5efa-4875-aad5-f512a727752f, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 834.814044] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Releasing lock "refresh_cache-1848c251-0a0c-4c36-8bd0-ff37befde2a0" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.814662] env[62692]: DEBUG nova.compute.manager [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 834.815217] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 834.816307] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14080142-f2aa-47ee-9976-2b4d71780ff2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.830202] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 834.830612] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aa0e8cfb-bced-406e-b51c-f8d1f2001db5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.839890] env[62692]: DEBUG oslo_vmware.api [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Waiting for the task: (returnval){ [ 834.839890] env[62692]: value = "task-1141114" [ 834.839890] env[62692]: _type = "Task" [ 834.839890] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.856741] env[62692]: DEBUG oslo_vmware.api [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141114, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.877197] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cbeebc8-0d5d-4daa-a5de-587e650633d1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.907531] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff824c7d-c377-450a-be34-ad90487585f9 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 834.911642] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ee4024a0-9587-468a-ae8c-1e110611967d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.920266] env[62692]: DEBUG oslo_vmware.api [None req-ff824c7d-c377-450a-be34-ad90487585f9 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 834.920266] env[62692]: value = "task-1141115" [ 834.920266] env[62692]: _type = "Task" [ 834.920266] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.929050] env[62692]: DEBUG oslo_vmware.api [None req-ff824c7d-c377-450a-be34-ad90487585f9 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141115, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.074596] env[62692]: DEBUG oslo_concurrency.lockutils [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "b1d6409a-2733-470f-a929-672fe1631b1b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.141974] env[62692]: DEBUG nova.compute.manager [req-53a591fb-98b3-4252-b64a-ffb8ac185491 req-876a4b05-e122-433d-b08e-2ddeb60ce6fd service nova] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Received event network-vif-plugged-4eb54953-e726-4bd7-9793-30a73f994d30 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 835.141974] env[62692]: DEBUG oslo_concurrency.lockutils [req-53a591fb-98b3-4252-b64a-ffb8ac185491 req-876a4b05-e122-433d-b08e-2ddeb60ce6fd service nova] Acquiring lock "8b75e781-930e-4885-81d2-8a8929d6c39b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.142209] env[62692]: DEBUG oslo_concurrency.lockutils [req-53a591fb-98b3-4252-b64a-ffb8ac185491 req-876a4b05-e122-433d-b08e-2ddeb60ce6fd service nova] Lock "8b75e781-930e-4885-81d2-8a8929d6c39b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.142377] env[62692]: DEBUG oslo_concurrency.lockutils [req-53a591fb-98b3-4252-b64a-ffb8ac185491 req-876a4b05-e122-433d-b08e-2ddeb60ce6fd service nova] Lock "8b75e781-930e-4885-81d2-8a8929d6c39b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.142547] env[62692]: DEBUG nova.compute.manager [req-53a591fb-98b3-4252-b64a-ffb8ac185491 req-876a4b05-e122-433d-b08e-2ddeb60ce6fd service nova] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] No waiting events found dispatching network-vif-plugged-4eb54953-e726-4bd7-9793-30a73f994d30 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 835.142717] env[62692]: WARNING nova.compute.manager [req-53a591fb-98b3-4252-b64a-ffb8ac185491 req-876a4b05-e122-433d-b08e-2ddeb60ce6fd service nova] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Received unexpected event network-vif-plugged-4eb54953-e726-4bd7-9793-30a73f994d30 for instance with vm_state building and task_state spawning. [ 835.162386] env[62692]: DEBUG oslo_concurrency.lockutils [None req-09002d2d-7e74-4699-8b3e-5d2429d1bef7 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "b1d6409a-2733-470f-a929-672fe1631b1b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.208s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.162644] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Task: {'id': task-1141112, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.164523] env[62692]: DEBUG oslo_concurrency.lockutils [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "b1d6409a-2733-470f-a929-672fe1631b1b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.090s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.164753] env[62692]: DEBUG oslo_concurrency.lockutils [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "b1d6409a-2733-470f-a929-672fe1631b1b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.165052] env[62692]: DEBUG oslo_concurrency.lockutils [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "b1d6409a-2733-470f-a929-672fe1631b1b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.165629] env[62692]: DEBUG oslo_concurrency.lockutils [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "b1d6409a-2733-470f-a929-672fe1631b1b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.167923] env[62692]: INFO nova.compute.manager [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Terminating instance [ 835.169972] env[62692]: DEBUG nova.compute.manager [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 835.170204] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 835.174049] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d8c2e6f-3aa7-4eb7-9f9b-cdca1d6284d0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.183689] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 835.183957] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d4df0051-8c5d-4976-b336-98647a495c07 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.191747] env[62692]: DEBUG oslo_vmware.api [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 835.191747] env[62692]: value = "task-1141116" [ 835.191747] env[62692]: _type = "Task" [ 835.191747] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.204613] env[62692]: DEBUG oslo_vmware.api [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141116, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.247288] env[62692]: DEBUG nova.network.neutron [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Successfully updated port: 4eb54953-e726-4bd7-9793-30a73f994d30 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 835.253620] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141113, 'name': ReconfigVM_Task, 'duration_secs': 0.436363} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.257345] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Reconfigured VM instance instance-00000031 to attach disk [datastore2] 7a3f9c7d-03cf-4177-88db-acf33cc03feb/7a3f9c7d-03cf-4177-88db-acf33cc03feb.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 835.258532] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-73793379-0bb4-4ca1-a14f-99f60c142d90 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.267615] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Waiting for the task: (returnval){ [ 835.267615] env[62692]: value = "task-1141117" [ 835.267615] env[62692]: _type = "Task" [ 835.267615] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.275023] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebd86cdd-74fb-4190-a972-1d0ab0b86994 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.285106] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141117, 'name': Rename_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.286247] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48a30fbf-ef22-41b7-b75e-48a6dfc80a24 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.321075] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6135547a-b1a9-4f0d-92a4-98773ca0913a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.329928] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-064fa1cf-e96b-4a71-8f97-cfdb9a0c645e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.346636] env[62692]: DEBUG nova.compute.provider_tree [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 835.359656] env[62692]: DEBUG oslo_vmware.api [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141114, 'name': PowerOffVM_Task, 'duration_secs': 0.169925} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.359656] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 835.359656] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 835.359656] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cfe391c4-45b0-40eb-8ee9-a11e8b39b0c6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.391603] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 835.392166] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 835.392166] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Deleting the datastore file [datastore1] 1848c251-0a0c-4c36-8bd0-ff37befde2a0 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 835.392419] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eca032b8-32a6-430e-90c4-f532a3b01ede {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.401555] env[62692]: DEBUG oslo_vmware.api [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Waiting for the task: (returnval){ [ 835.401555] env[62692]: value = "task-1141119" [ 835.401555] env[62692]: _type = "Task" [ 835.401555] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.411342] env[62692]: DEBUG oslo_vmware.api [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141119, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.431134] env[62692]: DEBUG oslo_vmware.api [None req-ff824c7d-c377-450a-be34-ad90487585f9 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141115, 'name': PowerOffVM_Task, 'duration_secs': 0.251826} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.431441] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff824c7d-c377-450a-be34-ad90487585f9 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 835.437309] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff824c7d-c377-450a-be34-ad90487585f9 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Reconfiguring VM instance instance-00000029 to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 835.437605] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8f655711-b32e-43cb-bf28-bd448555b9c6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.457473] env[62692]: DEBUG oslo_vmware.api [None req-ff824c7d-c377-450a-be34-ad90487585f9 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 835.457473] env[62692]: value = "task-1141120" [ 835.457473] env[62692]: _type = "Task" [ 835.457473] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.468115] env[62692]: DEBUG oslo_vmware.api [None req-ff824c7d-c377-450a-be34-ad90487585f9 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141120, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.658561] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Task: {'id': task-1141112, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.674275] env[62692]: DEBUG nova.compute.manager [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 835.702339] env[62692]: DEBUG oslo_vmware.api [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141116, 'name': PowerOffVM_Task, 'duration_secs': 0.343256} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.702659] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 835.702852] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 835.703204] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-23ea5367-7db5-4759-bfb1-d2f4a34d6dc2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.758180] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "refresh_cache-8b75e781-930e-4885-81d2-8a8929d6c39b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.758349] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquired lock "refresh_cache-8b75e781-930e-4885-81d2-8a8929d6c39b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.758533] env[62692]: DEBUG nova.network.neutron [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 835.778567] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141117, 'name': Rename_Task, 'duration_secs': 0.244141} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.778871] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 835.780440] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8f460972-c050-4f6f-9fbf-7a0ffc284b2f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.806630] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Waiting for the task: (returnval){ [ 835.806630] env[62692]: value = "task-1141122" [ 835.806630] env[62692]: _type = "Task" [ 835.806630] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.815516] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141122, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.819354] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 835.819693] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 835.819779] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Deleting the datastore file [datastore2] b1d6409a-2733-470f-a929-672fe1631b1b {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 835.820352] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f9e20439-dc95-4a33-ad1e-91c21b3d109e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.827574] env[62692]: DEBUG oslo_vmware.api [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 835.827574] env[62692]: value = "task-1141123" [ 835.827574] env[62692]: _type = "Task" [ 835.827574] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.837559] env[62692]: DEBUG oslo_vmware.api [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141123, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.895934] env[62692]: DEBUG nova.scheduler.client.report [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Updated inventory for provider 470758a8-5efa-4875-aad5-f512a727752f with generation 70 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 835.896091] env[62692]: DEBUG nova.compute.provider_tree [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Updating resource provider 470758a8-5efa-4875-aad5-f512a727752f generation from 70 to 71 during operation: update_inventory {{(pid=62692) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 835.896239] env[62692]: DEBUG nova.compute.provider_tree [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 835.912509] env[62692]: DEBUG oslo_vmware.api [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Task: {'id': task-1141119, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1363} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.912509] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 835.912509] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 835.912509] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 835.912509] env[62692]: INFO nova.compute.manager [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Took 1.10 seconds to destroy the instance on the hypervisor. [ 835.912831] env[62692]: DEBUG oslo.service.loopingcall [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 835.912831] env[62692]: DEBUG nova.compute.manager [-] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 835.912900] env[62692]: DEBUG nova.network.neutron [-] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 835.932501] env[62692]: DEBUG nova.network.neutron [-] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 835.967701] env[62692]: DEBUG oslo_vmware.api [None req-ff824c7d-c377-450a-be34-ad90487585f9 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141120, 'name': ReconfigVM_Task, 'duration_secs': 0.246027} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.967701] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff824c7d-c377-450a-be34-ad90487585f9 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Reconfigured VM instance instance-00000029 to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 835.967701] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff824c7d-c377-450a-be34-ad90487585f9 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 835.967949] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-13f39220-fbb0-4c21-a3df-a100e697a4f3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.975665] env[62692]: DEBUG oslo_vmware.api [None req-ff824c7d-c377-450a-be34-ad90487585f9 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 835.975665] env[62692]: value = "task-1141124" [ 835.975665] env[62692]: _type = "Task" [ 835.975665] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.985713] env[62692]: DEBUG oslo_vmware.api [None req-ff824c7d-c377-450a-be34-ad90487585f9 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141124, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.162153] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Task: {'id': task-1141112, 'name': ReconfigVM_Task, 'duration_secs': 1.623616} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.162153] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Reconfigured VM instance instance-00000033 to attach disk [datastore2] ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba/ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 836.162153] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fbf6050d-daba-4478-9d8f-fe917515c532 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.171325] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Waiting for the task: (returnval){ [ 836.171325] env[62692]: value = "task-1141125" [ 836.171325] env[62692]: _type = "Task" [ 836.171325] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.182383] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Task: {'id': task-1141125, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.201321] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.316957] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141122, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.318351] env[62692]: DEBUG nova.network.neutron [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 836.338829] env[62692]: DEBUG oslo_vmware.api [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141123, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.209814} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.339126] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 836.339317] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 836.339495] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 836.339668] env[62692]: INFO nova.compute.manager [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Took 1.17 seconds to destroy the instance on the hypervisor. [ 836.340016] env[62692]: DEBUG oslo.service.loopingcall [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 836.340275] env[62692]: DEBUG nova.compute.manager [-] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 836.340473] env[62692]: DEBUG nova.network.neutron [-] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 836.404273] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.986s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.404830] env[62692]: DEBUG nova.compute.manager [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 836.409192] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.680s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.409413] env[62692]: DEBUG nova.objects.instance [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Lazy-loading 'resources' on Instance uuid 6097d6b2-52d5-4765-94c0-d5f3609464d8 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 836.434804] env[62692]: DEBUG nova.network.neutron [-] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.489216] env[62692]: DEBUG oslo_vmware.api [None req-ff824c7d-c377-450a-be34-ad90487585f9 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141124, 'name': PowerOnVM_Task, 'duration_secs': 0.395007} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.490224] env[62692]: DEBUG nova.network.neutron [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Updating instance_info_cache with network_info: [{"id": "4eb54953-e726-4bd7-9793-30a73f994d30", "address": "fa:16:3e:6e:4f:28", "network": {"id": "d0e00428-1d04-4b63-8975-7f3085784358", "bridge": "br-int", "label": "tempest-ServersTestJSON-408035146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e13f8888d5e84eed8247101558e29a56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4eb54953-e7", "ovs_interfaceid": "4eb54953-e726-4bd7-9793-30a73f994d30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.491408] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff824c7d-c377-450a-be34-ad90487585f9 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 836.495027] env[62692]: DEBUG nova.compute.manager [None req-ff824c7d-c377-450a-be34-ad90487585f9 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 836.495027] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7771d8ca-221f-44eb-80f3-05dd2c7232ef {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.681912] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Task: {'id': task-1141125, 'name': Rename_Task, 'duration_secs': 0.157594} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.682244] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 836.682485] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-630800e4-1f9f-4b1a-9b0d-39d947ee12d0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.687224] env[62692]: DEBUG nova.compute.manager [req-d6e76d08-bda0-4a80-ba35-04e0ca629dab req-d92108b4-c1bd-440f-a85b-3a538e828317 service nova] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Received event network-vif-deleted-ce7100ab-22d6-4268-8f4c-844d9658c6d1 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 836.687418] env[62692]: INFO nova.compute.manager [req-d6e76d08-bda0-4a80-ba35-04e0ca629dab req-d92108b4-c1bd-440f-a85b-3a538e828317 service nova] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Neutron deleted interface ce7100ab-22d6-4268-8f4c-844d9658c6d1; detaching it from the instance and deleting it from the info cache [ 836.687607] env[62692]: DEBUG nova.network.neutron [req-d6e76d08-bda0-4a80-ba35-04e0ca629dab req-d92108b4-c1bd-440f-a85b-3a538e828317 service nova] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.690839] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Waiting for the task: (returnval){ [ 836.690839] env[62692]: value = "task-1141126" [ 836.690839] env[62692]: _type = "Task" [ 836.690839] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.703381] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Task: {'id': task-1141126, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.817393] env[62692]: DEBUG oslo_vmware.api [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141122, 'name': PowerOnVM_Task, 'duration_secs': 0.561399} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.819672] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 836.819672] env[62692]: DEBUG nova.compute.manager [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 836.819672] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0198948-6987-476f-99ae-1af4f1592ae6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.912823] env[62692]: DEBUG nova.compute.utils [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 836.920106] env[62692]: DEBUG nova.compute.manager [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 836.920106] env[62692]: DEBUG nova.network.neutron [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 836.937108] env[62692]: INFO nova.compute.manager [-] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Took 1.02 seconds to deallocate network for instance. [ 836.992711] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Releasing lock "refresh_cache-8b75e781-930e-4885-81d2-8a8929d6c39b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.993090] env[62692]: DEBUG nova.compute.manager [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Instance network_info: |[{"id": "4eb54953-e726-4bd7-9793-30a73f994d30", "address": "fa:16:3e:6e:4f:28", "network": {"id": "d0e00428-1d04-4b63-8975-7f3085784358", "bridge": "br-int", "label": "tempest-ServersTestJSON-408035146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e13f8888d5e84eed8247101558e29a56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4eb54953-e7", "ovs_interfaceid": "4eb54953-e726-4bd7-9793-30a73f994d30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 836.993496] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6e:4f:28', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4eb54953-e726-4bd7-9793-30a73f994d30', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 837.001509] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Creating folder: Project (e13f8888d5e84eed8247101558e29a56). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 837.005408] env[62692]: DEBUG nova.policy [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd33baeebac3d4fad907677f763bf71be', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '43057671f72b4bfa83f33855a1bf4422', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 837.007042] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4629fc06-c01e-4197-8d08-0152e6fc7d8d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.028224] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Created folder: Project (e13f8888d5e84eed8247101558e29a56) in parent group-v248868. [ 837.028224] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Creating folder: Instances. Parent ref: group-v248935. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 837.029674] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-41bef0e8-0fe7-4bd8-9ddf-d23f3e37e079 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.042302] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Created folder: Instances in parent group-v248935. [ 837.042571] env[62692]: DEBUG oslo.service.loopingcall [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 837.042775] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 837.043011] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d1866683-d23f-469a-8a50-43e3060d4e8c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.068346] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 837.068346] env[62692]: value = "task-1141129" [ 837.068346] env[62692]: _type = "Task" [ 837.068346] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.081116] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141129, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.161037] env[62692]: DEBUG nova.network.neutron [-] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.177524] env[62692]: DEBUG nova.compute.manager [req-c827d866-ad31-44a1-ad50-2ed714558a2c req-3880e079-db07-4b93-9616-2cc42b3cd981 service nova] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Received event network-changed-4eb54953-e726-4bd7-9793-30a73f994d30 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 837.177731] env[62692]: DEBUG nova.compute.manager [req-c827d866-ad31-44a1-ad50-2ed714558a2c req-3880e079-db07-4b93-9616-2cc42b3cd981 service nova] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Refreshing instance network info cache due to event network-changed-4eb54953-e726-4bd7-9793-30a73f994d30. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 837.177939] env[62692]: DEBUG oslo_concurrency.lockutils [req-c827d866-ad31-44a1-ad50-2ed714558a2c req-3880e079-db07-4b93-9616-2cc42b3cd981 service nova] Acquiring lock "refresh_cache-8b75e781-930e-4885-81d2-8a8929d6c39b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.178154] env[62692]: DEBUG oslo_concurrency.lockutils [req-c827d866-ad31-44a1-ad50-2ed714558a2c req-3880e079-db07-4b93-9616-2cc42b3cd981 service nova] Acquired lock "refresh_cache-8b75e781-930e-4885-81d2-8a8929d6c39b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.178245] env[62692]: DEBUG nova.network.neutron [req-c827d866-ad31-44a1-ad50-2ed714558a2c req-3880e079-db07-4b93-9616-2cc42b3cd981 service nova] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Refreshing network info cache for port 4eb54953-e726-4bd7-9793-30a73f994d30 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 837.190443] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-619f9d89-a066-4915-ad4a-157897b8d99a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.206930] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Task: {'id': task-1141126, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.213252] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44b2f680-e3f7-484b-aced-61f0cbc13178 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.252147] env[62692]: DEBUG nova.compute.manager [req-d6e76d08-bda0-4a80-ba35-04e0ca629dab req-d92108b4-c1bd-440f-a85b-3a538e828317 service nova] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Detach interface failed, port_id=ce7100ab-22d6-4268-8f4c-844d9658c6d1, reason: Instance b1d6409a-2733-470f-a929-672fe1631b1b could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 837.339777] env[62692]: DEBUG oslo_concurrency.lockutils [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.418581] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82f25894-9f23-4a68-930c-c0470825e184 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.423265] env[62692]: DEBUG nova.compute.manager [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 837.432343] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adc11140-ad88-4ad6-a9e5-141e2b6ca520 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.467513] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.469230] env[62692]: DEBUG nova.network.neutron [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Successfully created port: 951b0add-cfaa-4506-a6ac-56bcb8d1041e {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 837.471729] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cdb4221-34f1-4180-ad52-4ada7fb7f24d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.481991] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43bcf0c0-523a-4e53-b3e7-cdbe9fddd8d4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.500124] env[62692]: DEBUG nova.compute.provider_tree [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 837.580154] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141129, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.664410] env[62692]: INFO nova.compute.manager [-] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Took 1.32 seconds to deallocate network for instance. [ 837.712565] env[62692]: DEBUG oslo_vmware.api [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Task: {'id': task-1141126, 'name': PowerOnVM_Task, 'duration_secs': 0.55487} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.713853] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 837.714386] env[62692]: INFO nova.compute.manager [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Took 10.72 seconds to spawn the instance on the hypervisor. [ 837.714521] env[62692]: DEBUG nova.compute.manager [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 837.716136] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1897084-33c4-448c-b386-bede4a78bb73 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.042989] env[62692]: DEBUG nova.scheduler.client.report [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Updated inventory for provider 470758a8-5efa-4875-aad5-f512a727752f with generation 71 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 838.043280] env[62692]: DEBUG nova.compute.provider_tree [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Updating resource provider 470758a8-5efa-4875-aad5-f512a727752f generation from 71 to 72 during operation: update_inventory {{(pid=62692) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 838.043457] env[62692]: DEBUG nova.compute.provider_tree [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 838.079921] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141129, 'name': CreateVM_Task, 'duration_secs': 0.587289} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.080120] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 838.080814] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.080982] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.081356] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 838.081622] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b18ec527-3342-41c2-b679-9b6a6c5a8ac7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.087543] env[62692]: DEBUG oslo_vmware.api [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 838.087543] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52f3ebe0-6d84-4f12-a035-07124a43c570" [ 838.087543] env[62692]: _type = "Task" [ 838.087543] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.096109] env[62692]: DEBUG oslo_vmware.api [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52f3ebe0-6d84-4f12-a035-07124a43c570, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.173341] env[62692]: DEBUG oslo_concurrency.lockutils [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.180491] env[62692]: DEBUG nova.network.neutron [req-c827d866-ad31-44a1-ad50-2ed714558a2c req-3880e079-db07-4b93-9616-2cc42b3cd981 service nova] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Updated VIF entry in instance network info cache for port 4eb54953-e726-4bd7-9793-30a73f994d30. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 838.180898] env[62692]: DEBUG nova.network.neutron [req-c827d866-ad31-44a1-ad50-2ed714558a2c req-3880e079-db07-4b93-9616-2cc42b3cd981 service nova] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Updating instance_info_cache with network_info: [{"id": "4eb54953-e726-4bd7-9793-30a73f994d30", "address": "fa:16:3e:6e:4f:28", "network": {"id": "d0e00428-1d04-4b63-8975-7f3085784358", "bridge": "br-int", "label": "tempest-ServersTestJSON-408035146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e13f8888d5e84eed8247101558e29a56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4eb54953-e7", "ovs_interfaceid": "4eb54953-e726-4bd7-9793-30a73f994d30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.246306] env[62692]: INFO nova.compute.manager [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Took 29.41 seconds to build instance. [ 838.306751] env[62692]: INFO nova.compute.manager [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Rebuilding instance [ 838.353696] env[62692]: DEBUG nova.compute.manager [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 838.355649] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-599f824e-76d7-4751-b2c4-bc06ac9ebb46 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.434484] env[62692]: DEBUG nova.compute.manager [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 838.462264] env[62692]: DEBUG nova.virt.hardware [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:32:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='a53d9694-cbf0-4c37-a7c0-aa52b2363bc3',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1692421942',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 838.462614] env[62692]: DEBUG nova.virt.hardware [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 838.462836] env[62692]: DEBUG nova.virt.hardware [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 838.463104] env[62692]: DEBUG nova.virt.hardware [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 838.463325] env[62692]: DEBUG nova.virt.hardware [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 838.463573] env[62692]: DEBUG nova.virt.hardware [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 838.463822] env[62692]: DEBUG nova.virt.hardware [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 838.464380] env[62692]: DEBUG nova.virt.hardware [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 838.464380] env[62692]: DEBUG nova.virt.hardware [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 838.464553] env[62692]: DEBUG nova.virt.hardware [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 838.464783] env[62692]: DEBUG nova.virt.hardware [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 838.465695] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eecfb663-c139-4c58-b63f-993cb5524d9b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.475342] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-054eccc2-755f-4f83-af71-5670cde7fc64 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.548571] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.139s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.552270] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.962s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.553754] env[62692]: INFO nova.compute.claims [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 838.571891] env[62692]: INFO nova.scheduler.client.report [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Deleted allocations for instance 6097d6b2-52d5-4765-94c0-d5f3609464d8 [ 838.601410] env[62692]: DEBUG oslo_vmware.api [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52f3ebe0-6d84-4f12-a035-07124a43c570, 'name': SearchDatastore_Task, 'duration_secs': 0.034853} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.601725] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.601967] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 838.602464] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.602613] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.602849] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 838.604798] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-951672d0-b1a4-4380-8b39-5572cd0ddbe3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.614235] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 838.614385] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 838.618382] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8e4a764-9444-4288-a492-246014a02969 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.622624] env[62692]: DEBUG oslo_vmware.api [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 838.622624] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52ef1197-cbf7-6302-10ef-16acdabf7720" [ 838.622624] env[62692]: _type = "Task" [ 838.622624] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.633198] env[62692]: DEBUG oslo_vmware.api [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52ef1197-cbf7-6302-10ef-16acdabf7720, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.685253] env[62692]: DEBUG oslo_concurrency.lockutils [req-c827d866-ad31-44a1-ad50-2ed714558a2c req-3880e079-db07-4b93-9616-2cc42b3cd981 service nova] Releasing lock "refresh_cache-8b75e781-930e-4885-81d2-8a8929d6c39b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.750876] env[62692]: DEBUG oslo_concurrency.lockutils [None req-da142d8a-3b84-440f-b047-99df46935b90 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Lock "ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.944s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.869920] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 838.873376] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1c67c02d-6083-4b9c-910f-80846d5bd07d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.881641] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Waiting for the task: (returnval){ [ 838.881641] env[62692]: value = "task-1141130" [ 838.881641] env[62692]: _type = "Task" [ 838.881641] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.892132] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Task: {'id': task-1141130, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.081154] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3e9b414e-941f-46dc-b295-023465859df6 tempest-ServerAddressesNegativeTestJSON-1441386536 tempest-ServerAddressesNegativeTestJSON-1441386536-project-member] Lock "6097d6b2-52d5-4765-94c0-d5f3609464d8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.435s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.086720] env[62692]: DEBUG nova.compute.manager [req-c28bd8e0-2ed7-440d-b5ad-0d3bbd5391f0 req-5ea685a2-0dfb-49d1-8c19-ac2f748ef542 service nova] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Received event network-vif-plugged-951b0add-cfaa-4506-a6ac-56bcb8d1041e {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 839.086996] env[62692]: DEBUG oslo_concurrency.lockutils [req-c28bd8e0-2ed7-440d-b5ad-0d3bbd5391f0 req-5ea685a2-0dfb-49d1-8c19-ac2f748ef542 service nova] Acquiring lock "2ccb76cf-f641-4306-a137-fb2417285df9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.087164] env[62692]: DEBUG oslo_concurrency.lockutils [req-c28bd8e0-2ed7-440d-b5ad-0d3bbd5391f0 req-5ea685a2-0dfb-49d1-8c19-ac2f748ef542 service nova] Lock "2ccb76cf-f641-4306-a137-fb2417285df9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.087364] env[62692]: DEBUG oslo_concurrency.lockutils [req-c28bd8e0-2ed7-440d-b5ad-0d3bbd5391f0 req-5ea685a2-0dfb-49d1-8c19-ac2f748ef542 service nova] Lock "2ccb76cf-f641-4306-a137-fb2417285df9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.088047] env[62692]: DEBUG nova.compute.manager [req-c28bd8e0-2ed7-440d-b5ad-0d3bbd5391f0 req-5ea685a2-0dfb-49d1-8c19-ac2f748ef542 service nova] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] No waiting events found dispatching network-vif-plugged-951b0add-cfaa-4506-a6ac-56bcb8d1041e {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 839.088047] env[62692]: WARNING nova.compute.manager [req-c28bd8e0-2ed7-440d-b5ad-0d3bbd5391f0 req-5ea685a2-0dfb-49d1-8c19-ac2f748ef542 service nova] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Received unexpected event network-vif-plugged-951b0add-cfaa-4506-a6ac-56bcb8d1041e for instance with vm_state building and task_state spawning. [ 839.137780] env[62692]: DEBUG oslo_vmware.api [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52ef1197-cbf7-6302-10ef-16acdabf7720, 'name': SearchDatastore_Task, 'duration_secs': 0.011228} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.138503] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ece40cc8-244f-4b0a-926d-c2e082016088 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.149732] env[62692]: DEBUG oslo_vmware.api [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 839.149732] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]521d4142-fc51-3a1b-30b5-ff6182b41a88" [ 839.149732] env[62692]: _type = "Task" [ 839.149732] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.150526] env[62692]: DEBUG nova.network.neutron [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Successfully updated port: 951b0add-cfaa-4506-a6ac-56bcb8d1041e {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 839.169903] env[62692]: DEBUG oslo_vmware.api [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]521d4142-fc51-3a1b-30b5-ff6182b41a88, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.179242] env[62692]: DEBUG nova.compute.manager [req-76fbd4eb-0f8c-4bf8-81c6-15991bc09265 req-370d4665-40f7-436b-acf9-e54e7e74aadb service nova] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Received event network-changed-951b0add-cfaa-4506-a6ac-56bcb8d1041e {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 839.179437] env[62692]: DEBUG nova.compute.manager [req-76fbd4eb-0f8c-4bf8-81c6-15991bc09265 req-370d4665-40f7-436b-acf9-e54e7e74aadb service nova] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Refreshing instance network info cache due to event network-changed-951b0add-cfaa-4506-a6ac-56bcb8d1041e. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 839.179645] env[62692]: DEBUG oslo_concurrency.lockutils [req-76fbd4eb-0f8c-4bf8-81c6-15991bc09265 req-370d4665-40f7-436b-acf9-e54e7e74aadb service nova] Acquiring lock "refresh_cache-2ccb76cf-f641-4306-a137-fb2417285df9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.179786] env[62692]: DEBUG oslo_concurrency.lockutils [req-76fbd4eb-0f8c-4bf8-81c6-15991bc09265 req-370d4665-40f7-436b-acf9-e54e7e74aadb service nova] Acquired lock "refresh_cache-2ccb76cf-f641-4306-a137-fb2417285df9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.179988] env[62692]: DEBUG nova.network.neutron [req-76fbd4eb-0f8c-4bf8-81c6-15991bc09265 req-370d4665-40f7-436b-acf9-e54e7e74aadb service nova] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Refreshing network info cache for port 951b0add-cfaa-4506-a6ac-56bcb8d1041e {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 839.255288] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Acquiring lock "ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.255560] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Lock "ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.255773] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Acquiring lock "ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.255952] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Lock "ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.256171] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Lock "ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.258034] env[62692]: DEBUG nova.compute.manager [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 839.261863] env[62692]: INFO nova.compute.manager [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Terminating instance [ 839.264796] env[62692]: DEBUG nova.compute.manager [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 839.265071] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 839.266530] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2d1020b-b382-46d2-85ab-07707157df49 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.275664] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 839.275772] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5e0e7ddf-3606-4a7a-955b-8cb400d2d650 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.283196] env[62692]: DEBUG oslo_vmware.api [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Waiting for the task: (returnval){ [ 839.283196] env[62692]: value = "task-1141131" [ 839.283196] env[62692]: _type = "Task" [ 839.283196] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.297207] env[62692]: DEBUG oslo_vmware.api [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Task: {'id': task-1141131, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.392674] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Task: {'id': task-1141130, 'name': PowerOffVM_Task, 'duration_secs': 0.290911} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.393016] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 839.393307] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 839.394216] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a7e5818-10d5-4882-913c-ad2e23321225 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.404803] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 839.404803] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cd5f04b0-5327-4b38-9640-c1f1927cd9b7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.443142] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 839.443550] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 839.443759] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Deleting the datastore file [datastore2] 7a3f9c7d-03cf-4177-88db-acf33cc03feb {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 839.444696] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-45bdb747-9978-4ed8-9a58-0e30f4188d3e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.453629] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Waiting for the task: (returnval){ [ 839.453629] env[62692]: value = "task-1141133" [ 839.453629] env[62692]: _type = "Task" [ 839.453629] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.464147] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Task: {'id': task-1141133, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.549893] env[62692]: DEBUG oslo_concurrency.lockutils [None req-010bcb5f-4168-4c3d-88d0-9f094c7e6b68 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Acquiring lock "abee88ff-5cf5-4bf1-91e0-93b19cf30046" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.550253] env[62692]: DEBUG oslo_concurrency.lockutils [None req-010bcb5f-4168-4c3d-88d0-9f094c7e6b68 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lock "abee88ff-5cf5-4bf1-91e0-93b19cf30046" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.668018] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquiring lock "refresh_cache-2ccb76cf-f641-4306-a137-fb2417285df9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.669100] env[62692]: DEBUG oslo_vmware.api [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]521d4142-fc51-3a1b-30b5-ff6182b41a88, 'name': SearchDatastore_Task, 'duration_secs': 0.026462} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.669345] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.670248] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 8b75e781-930e-4885-81d2-8a8929d6c39b/8b75e781-930e-4885-81d2-8a8929d6c39b.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 839.670248] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-770c1fd9-022c-4767-98ac-feeab32a750e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.677712] env[62692]: DEBUG oslo_vmware.api [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 839.677712] env[62692]: value = "task-1141134" [ 839.677712] env[62692]: _type = "Task" [ 839.677712] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.688319] env[62692]: DEBUG oslo_vmware.api [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141134, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.723025] env[62692]: DEBUG nova.network.neutron [req-76fbd4eb-0f8c-4bf8-81c6-15991bc09265 req-370d4665-40f7-436b-acf9-e54e7e74aadb service nova] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 839.773362] env[62692]: DEBUG nova.network.neutron [req-76fbd4eb-0f8c-4bf8-81c6-15991bc09265 req-370d4665-40f7-436b-acf9-e54e7e74aadb service nova] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.778908] env[62692]: DEBUG oslo_concurrency.lockutils [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.804218] env[62692]: DEBUG oslo_vmware.api [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Task: {'id': task-1141131, 'name': PowerOffVM_Task, 'duration_secs': 0.208802} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.804487] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 839.804659] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 839.804912] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b6fbf42c-aea9-4d35-a746-941d751e1e58 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.865960] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 839.866219] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 839.866445] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Deleting the datastore file [datastore2] ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 839.866720] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-31d4d681-0611-4221-b8f7-094ca1e1615f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.874406] env[62692]: DEBUG oslo_vmware.api [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Waiting for the task: (returnval){ [ 839.874406] env[62692]: value = "task-1141136" [ 839.874406] env[62692]: _type = "Task" [ 839.874406] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.883309] env[62692]: DEBUG oslo_vmware.api [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Task: {'id': task-1141136, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.933754] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-293c9161-0488-4f3d-9056-50d1b0642aa2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.945860] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f38062e7-71cf-4c5f-85b4-c4cde1acc428 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.985065] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fd0fcbf-5282-4e68-a38c-a1abb5e59b53 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.997420] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccecd621-af83-418b-a5e7-dff78c4a3250 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.003387] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Task: {'id': task-1141133, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177632} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.003387] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 840.003694] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 840.004099] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 840.018028] env[62692]: DEBUG nova.compute.provider_tree [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 840.054385] env[62692]: DEBUG nova.compute.utils [None req-010bcb5f-4168-4c3d-88d0-9f094c7e6b68 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 840.189916] env[62692]: DEBUG oslo_vmware.api [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141134, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.276231] env[62692]: DEBUG oslo_concurrency.lockutils [req-76fbd4eb-0f8c-4bf8-81c6-15991bc09265 req-370d4665-40f7-436b-acf9-e54e7e74aadb service nova] Releasing lock "refresh_cache-2ccb76cf-f641-4306-a137-fb2417285df9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.276772] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquired lock "refresh_cache-2ccb76cf-f641-4306-a137-fb2417285df9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.277055] env[62692]: DEBUG nova.network.neutron [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 840.386355] env[62692]: DEBUG oslo_vmware.api [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Task: {'id': task-1141136, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.171718} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.386610] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 840.386800] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 840.386963] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 840.387144] env[62692]: INFO nova.compute.manager [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Took 1.12 seconds to destroy the instance on the hypervisor. [ 840.387382] env[62692]: DEBUG oslo.service.loopingcall [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 840.387567] env[62692]: DEBUG nova.compute.manager [-] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 840.387662] env[62692]: DEBUG nova.network.neutron [-] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 840.520839] env[62692]: DEBUG nova.scheduler.client.report [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 840.558447] env[62692]: DEBUG oslo_concurrency.lockutils [None req-010bcb5f-4168-4c3d-88d0-9f094c7e6b68 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lock "abee88ff-5cf5-4bf1-91e0-93b19cf30046" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.611152] env[62692]: DEBUG oslo_concurrency.lockutils [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.611426] env[62692]: DEBUG oslo_concurrency.lockutils [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.692243] env[62692]: DEBUG oslo_vmware.api [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141134, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.749758} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.692582] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 8b75e781-930e-4885-81d2-8a8929d6c39b/8b75e781-930e-4885-81d2-8a8929d6c39b.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 840.692803] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 840.693092] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6236addc-79b0-4392-9378-6f26efaa337e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.700978] env[62692]: DEBUG oslo_vmware.api [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 840.700978] env[62692]: value = "task-1141137" [ 840.700978] env[62692]: _type = "Task" [ 840.700978] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.709795] env[62692]: DEBUG oslo_vmware.api [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141137, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.831528] env[62692]: DEBUG nova.network.neutron [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 841.013487] env[62692]: DEBUG nova.network.neutron [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Updating instance_info_cache with network_info: [{"id": "951b0add-cfaa-4506-a6ac-56bcb8d1041e", "address": "fa:16:3e:70:90:85", "network": {"id": "a74c4d30-3568-4e50-98f5-d2fcc05ec728", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.89", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "989968e8796a4a70a4dbd045418a475f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap951b0add-cf", "ovs_interfaceid": "951b0add-cfaa-4506-a6ac-56bcb8d1041e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.025794] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.474s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.026320] env[62692]: DEBUG nova.compute.manager [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 841.029983] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.748s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.031713] env[62692]: INFO nova.compute.claims [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 841.062580] env[62692]: DEBUG nova.virt.hardware [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 841.063481] env[62692]: DEBUG nova.virt.hardware [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 841.063661] env[62692]: DEBUG nova.virt.hardware [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 841.066386] env[62692]: DEBUG nova.virt.hardware [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 841.066386] env[62692]: DEBUG nova.virt.hardware [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 841.066386] env[62692]: DEBUG nova.virt.hardware [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 841.066386] env[62692]: DEBUG nova.virt.hardware [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 841.066386] env[62692]: DEBUG nova.virt.hardware [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 841.066386] env[62692]: DEBUG nova.virt.hardware [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 841.066386] env[62692]: DEBUG nova.virt.hardware [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 841.066386] env[62692]: DEBUG nova.virt.hardware [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 841.067247] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7244268b-1b8a-49e9-96f0-9f1e5b6db3c3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.079903] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48b0d850-f04e-4cf7-a86b-c7540faebe20 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.097643] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Instance VIF info [] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 841.103655] env[62692]: DEBUG oslo.service.loopingcall [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 841.104504] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 841.104922] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3790d60e-8706-4e9c-ab24-fe39d8e35ed9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.124703] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 841.124703] env[62692]: value = "task-1141138" [ 841.124703] env[62692]: _type = "Task" [ 841.124703] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.134115] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141138, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.213439] env[62692]: DEBUG oslo_vmware.api [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141137, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084633} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.213728] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 841.214552] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc144b0c-585f-47e8-987e-b8a14e41140c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.238903] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Reconfiguring VM instance instance-00000034 to attach disk [datastore1] 8b75e781-930e-4885-81d2-8a8929d6c39b/8b75e781-930e-4885-81d2-8a8929d6c39b.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 841.239630] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8f5c6720-2362-493e-8b5c-a3c558dc48a7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.264681] env[62692]: DEBUG oslo_vmware.api [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 841.264681] env[62692]: value = "task-1141139" [ 841.264681] env[62692]: _type = "Task" [ 841.264681] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.274901] env[62692]: DEBUG oslo_vmware.api [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141139, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.315166] env[62692]: DEBUG nova.network.neutron [-] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.323159] env[62692]: DEBUG nova.compute.manager [req-033dcbf9-dd72-4931-b547-5b4763401717 req-ed6bf08b-d464-436b-8d31-b941c5f688ef service nova] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Received event network-vif-deleted-d996a5d5-fb43-4af0-9531-397082ddca3f {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 841.323159] env[62692]: INFO nova.compute.manager [req-033dcbf9-dd72-4931-b547-5b4763401717 req-ed6bf08b-d464-436b-8d31-b941c5f688ef service nova] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Neutron deleted interface d996a5d5-fb43-4af0-9531-397082ddca3f; detaching it from the instance and deleting it from the info cache [ 841.323159] env[62692]: DEBUG nova.network.neutron [req-033dcbf9-dd72-4931-b547-5b4763401717 req-ed6bf08b-d464-436b-8d31-b941c5f688ef service nova] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.519110] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Releasing lock "refresh_cache-2ccb76cf-f641-4306-a137-fb2417285df9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.519468] env[62692]: DEBUG nova.compute.manager [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Instance network_info: |[{"id": "951b0add-cfaa-4506-a6ac-56bcb8d1041e", "address": "fa:16:3e:70:90:85", "network": {"id": "a74c4d30-3568-4e50-98f5-d2fcc05ec728", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.89", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "989968e8796a4a70a4dbd045418a475f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap951b0add-cf", "ovs_interfaceid": "951b0add-cfaa-4506-a6ac-56bcb8d1041e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 841.519963] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:70:90:85', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd8383707-f093-40a7-a5ba-31b0e07cac45', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '951b0add-cfaa-4506-a6ac-56bcb8d1041e', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 841.530562] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Creating folder: Project (43057671f72b4bfa83f33855a1bf4422). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 841.532111] env[62692]: DEBUG nova.compute.utils [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 841.532926] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6d4b0cb3-0b34-48cd-b7ab-3129be8a980b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.535336] env[62692]: DEBUG nova.compute.manager [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 841.538249] env[62692]: DEBUG nova.network.neutron [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 841.556840] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Created folder: Project (43057671f72b4bfa83f33855a1bf4422) in parent group-v248868. [ 841.557093] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Creating folder: Instances. Parent ref: group-v248939. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 841.557762] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9d6f936f-8452-45f5-b615-f04cc8a5c9d5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.571730] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Created folder: Instances in parent group-v248939. [ 841.572022] env[62692]: DEBUG oslo.service.loopingcall [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 841.572213] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 841.572422] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9262fcdf-199b-422e-9a85-e8a414933dd8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.602605] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 841.602605] env[62692]: value = "task-1141142" [ 841.602605] env[62692]: _type = "Task" [ 841.602605] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.612953] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141142, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.637838] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141138, 'name': CreateVM_Task, 'duration_secs': 0.322208} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.639238] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 841.639238] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.639512] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.640031] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 841.640342] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f7b71bf-8098-4fb0-9d4d-22969d09cef3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.647336] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Waiting for the task: (returnval){ [ 841.647336] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]527a926d-e5c1-9b24-9bb2-7a996a90ef75" [ 841.647336] env[62692]: _type = "Task" [ 841.647336] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.653652] env[62692]: DEBUG nova.policy [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7d935dcee1d74802b3d04083cabc10a0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ef45ad1482dc4f5db51a5badc3cdce22', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 841.660344] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]527a926d-e5c1-9b24-9bb2-7a996a90ef75, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.669222] env[62692]: DEBUG oslo_concurrency.lockutils [None req-010bcb5f-4168-4c3d-88d0-9f094c7e6b68 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Acquiring lock "abee88ff-5cf5-4bf1-91e0-93b19cf30046" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.669498] env[62692]: DEBUG oslo_concurrency.lockutils [None req-010bcb5f-4168-4c3d-88d0-9f094c7e6b68 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lock "abee88ff-5cf5-4bf1-91e0-93b19cf30046" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.669780] env[62692]: INFO nova.compute.manager [None req-010bcb5f-4168-4c3d-88d0-9f094c7e6b68 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Attaching volume eccf2e07-c766-49a3-9929-1c9805b08d2a to /dev/sdb [ 841.709510] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-229e48b1-947b-4cca-88e7-c21039fc370c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.719113] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-660ff0e6-373f-4e0e-acf7-7d2b0c12fe1d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.739662] env[62692]: DEBUG nova.virt.block_device [None req-010bcb5f-4168-4c3d-88d0-9f094c7e6b68 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Updating existing volume attachment record: f2bc3ea5-f516-4abb-8f1f-79a3aa113aad {{(pid=62692) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 841.784031] env[62692]: DEBUG oslo_vmware.api [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141139, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.818954] env[62692]: INFO nova.compute.manager [-] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Took 1.43 seconds to deallocate network for instance. [ 841.826351] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fb1a542f-4a34-47f5-938f-91dcfe5bcb14 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.842436] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75db2e52-c573-45f6-a97a-c3ca3bab9355 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.883349] env[62692]: DEBUG nova.compute.manager [req-033dcbf9-dd72-4931-b547-5b4763401717 req-ed6bf08b-d464-436b-8d31-b941c5f688ef service nova] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Detach interface failed, port_id=d996a5d5-fb43-4af0-9531-397082ddca3f, reason: Instance ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 842.047026] env[62692]: DEBUG nova.compute.manager [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 842.121052] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141142, 'name': CreateVM_Task, 'duration_secs': 0.388074} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.121239] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 842.122075] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.122604] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.122604] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 842.123185] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4cdffa0f-575b-4583-ba26-8e2324755260 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.129673] env[62692]: DEBUG oslo_vmware.api [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Waiting for the task: (returnval){ [ 842.129673] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5261efeb-f977-2db4-7d3f-07c4673dc220" [ 842.129673] env[62692]: _type = "Task" [ 842.129673] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.138961] env[62692]: DEBUG oslo_vmware.api [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5261efeb-f977-2db4-7d3f-07c4673dc220, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.159537] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]527a926d-e5c1-9b24-9bb2-7a996a90ef75, 'name': SearchDatastore_Task, 'duration_secs': 0.016781} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.163017] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.163017] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 842.163017] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.163017] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.163017] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 842.165090] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d9a59e9c-d252-45d8-acf9-70f3c3259ebf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.168242] env[62692]: DEBUG nova.network.neutron [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Successfully created port: fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 842.181668] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 842.181668] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 842.182539] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b8b8487-e9fe-42ef-adf1-ef51a0b65122 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.189410] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Waiting for the task: (returnval){ [ 842.189410] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5273adb3-97a2-8466-5233-5cbf46dca11b" [ 842.189410] env[62692]: _type = "Task" [ 842.189410] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.201277] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5273adb3-97a2-8466-5233-5cbf46dca11b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.277106] env[62692]: DEBUG oslo_vmware.api [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141139, 'name': ReconfigVM_Task, 'duration_secs': 0.577899} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.280169] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Reconfigured VM instance instance-00000034 to attach disk [datastore1] 8b75e781-930e-4885-81d2-8a8929d6c39b/8b75e781-930e-4885-81d2-8a8929d6c39b.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 842.280958] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dc5b1368-3b17-4361-bd85-3fa463a8763d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.291531] env[62692]: DEBUG oslo_vmware.api [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 842.291531] env[62692]: value = "task-1141146" [ 842.291531] env[62692]: _type = "Task" [ 842.291531] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.307152] env[62692]: DEBUG oslo_vmware.api [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141146, 'name': Rename_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.327923] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.513310] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9113a85a-0c8c-4f9d-9a69-1729d6f89f40 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.525261] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fe05298-0ff1-446d-9392-fe50eee6d711 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.563294] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9db08ea-5c7f-4c13-afa8-f7f89fa06120 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.572781] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97ee1f0c-47e9-49f9-b2eb-0100e7c15257 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.595830] env[62692]: DEBUG nova.compute.provider_tree [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.640476] env[62692]: DEBUG oslo_vmware.api [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5261efeb-f977-2db4-7d3f-07c4673dc220, 'name': SearchDatastore_Task, 'duration_secs': 0.012811} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.640782] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.641088] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 842.641259] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.642062] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.642062] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 842.642062] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2fe65cea-7eed-4551-ac7a-382e945016af {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.653516] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 842.653796] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 842.654487] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ea8d8fc-7e52-453d-8ef7-6568c7a15d8f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.660848] env[62692]: DEBUG oslo_vmware.api [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Waiting for the task: (returnval){ [ 842.660848] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52df3e14-ab5e-088b-ca62-6e165d420b38" [ 842.660848] env[62692]: _type = "Task" [ 842.660848] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.668966] env[62692]: DEBUG oslo_vmware.api [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52df3e14-ab5e-088b-ca62-6e165d420b38, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.698942] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5273adb3-97a2-8466-5233-5cbf46dca11b, 'name': SearchDatastore_Task, 'duration_secs': 0.010782} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.699824] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c66c22be-21cf-4b6c-81d2-aee3766857ec {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.705013] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Waiting for the task: (returnval){ [ 842.705013] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]528e8e93-52f0-5d13-51af-91b788224891" [ 842.705013] env[62692]: _type = "Task" [ 842.705013] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.712723] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]528e8e93-52f0-5d13-51af-91b788224891, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.804635] env[62692]: DEBUG oslo_vmware.api [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141146, 'name': Rename_Task, 'duration_secs': 0.16123} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.804937] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 842.805373] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-05276a1f-2b5a-4949-839e-bc6f6a03a7d6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.812186] env[62692]: DEBUG oslo_vmware.api [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 842.812186] env[62692]: value = "task-1141147" [ 842.812186] env[62692]: _type = "Task" [ 842.812186] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.820158] env[62692]: DEBUG oslo_vmware.api [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141147, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.067566] env[62692]: DEBUG nova.compute.manager [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 843.092794] env[62692]: DEBUG nova.virt.hardware [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 843.093300] env[62692]: DEBUG nova.virt.hardware [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 843.093300] env[62692]: DEBUG nova.virt.hardware [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 843.093480] env[62692]: DEBUG nova.virt.hardware [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 843.093525] env[62692]: DEBUG nova.virt.hardware [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 843.093796] env[62692]: DEBUG nova.virt.hardware [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 843.093866] env[62692]: DEBUG nova.virt.hardware [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 843.094058] env[62692]: DEBUG nova.virt.hardware [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 843.094237] env[62692]: DEBUG nova.virt.hardware [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 843.094400] env[62692]: DEBUG nova.virt.hardware [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 843.094570] env[62692]: DEBUG nova.virt.hardware [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 843.095588] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a551ecde-1c59-4d20-ab29-83c42a443bff {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.098823] env[62692]: DEBUG nova.scheduler.client.report [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 843.108130] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efec765a-d99f-4ed9-aeeb-bfb42b5eb6b1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.171026] env[62692]: DEBUG oslo_vmware.api [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52df3e14-ab5e-088b-ca62-6e165d420b38, 'name': SearchDatastore_Task, 'duration_secs': 0.035016} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.171656] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ccec6fd-6e5b-44d2-b559-2eb0454f2852 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.177180] env[62692]: DEBUG oslo_vmware.api [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Waiting for the task: (returnval){ [ 843.177180] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]524f254d-7da2-f5be-0a1a-64456afda17a" [ 843.177180] env[62692]: _type = "Task" [ 843.177180] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.185059] env[62692]: DEBUG oslo_vmware.api [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]524f254d-7da2-f5be-0a1a-64456afda17a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.215952] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]528e8e93-52f0-5d13-51af-91b788224891, 'name': SearchDatastore_Task, 'duration_secs': 0.020925} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.216249] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.216519] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 7a3f9c7d-03cf-4177-88db-acf33cc03feb/7a3f9c7d-03cf-4177-88db-acf33cc03feb.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 843.216837] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-85f86080-0fad-4e27-b577-5f1f67bdef39 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.223474] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Waiting for the task: (returnval){ [ 843.223474] env[62692]: value = "task-1141148" [ 843.223474] env[62692]: _type = "Task" [ 843.223474] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.231690] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Task: {'id': task-1141148, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.329026] env[62692]: DEBUG oslo_vmware.api [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141147, 'name': PowerOnVM_Task, 'duration_secs': 0.468436} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.329026] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 843.329026] env[62692]: INFO nova.compute.manager [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Took 8.86 seconds to spawn the instance on the hypervisor. [ 843.329026] env[62692]: DEBUG nova.compute.manager [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 843.329026] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-030ca1a9-8046-4481-9a90-6237ffe1d96e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.604737] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.575s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.605317] env[62692]: DEBUG nova.compute.manager [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 843.609508] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.374s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.610545] env[62692]: INFO nova.compute.claims [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 843.689066] env[62692]: DEBUG oslo_vmware.api [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]524f254d-7da2-f5be-0a1a-64456afda17a, 'name': SearchDatastore_Task, 'duration_secs': 0.023459} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.689375] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.689646] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 2ccb76cf-f641-4306-a137-fb2417285df9/2ccb76cf-f641-4306-a137-fb2417285df9.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 843.690289] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-64023f54-2e55-401f-8709-6fdcdf543473 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.699100] env[62692]: DEBUG oslo_vmware.api [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Waiting for the task: (returnval){ [ 843.699100] env[62692]: value = "task-1141149" [ 843.699100] env[62692]: _type = "Task" [ 843.699100] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.708210] env[62692]: DEBUG oslo_vmware.api [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141149, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.736349] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Task: {'id': task-1141148, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.837408] env[62692]: DEBUG nova.compute.manager [req-59ce248b-d067-4155-901f-e67ea91eeb2b req-c50fe925-b5b0-47e8-af27-735650fffc7a service nova] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Received event network-vif-plugged-fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 843.837408] env[62692]: DEBUG oslo_concurrency.lockutils [req-59ce248b-d067-4155-901f-e67ea91eeb2b req-c50fe925-b5b0-47e8-af27-735650fffc7a service nova] Acquiring lock "c6a965d3-1c52-4f95-a226-9d15b7197ce7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.837408] env[62692]: DEBUG oslo_concurrency.lockutils [req-59ce248b-d067-4155-901f-e67ea91eeb2b req-c50fe925-b5b0-47e8-af27-735650fffc7a service nova] Lock "c6a965d3-1c52-4f95-a226-9d15b7197ce7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.837408] env[62692]: DEBUG oslo_concurrency.lockutils [req-59ce248b-d067-4155-901f-e67ea91eeb2b req-c50fe925-b5b0-47e8-af27-735650fffc7a service nova] Lock "c6a965d3-1c52-4f95-a226-9d15b7197ce7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.837408] env[62692]: DEBUG nova.compute.manager [req-59ce248b-d067-4155-901f-e67ea91eeb2b req-c50fe925-b5b0-47e8-af27-735650fffc7a service nova] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] No waiting events found dispatching network-vif-plugged-fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 843.837836] env[62692]: WARNING nova.compute.manager [req-59ce248b-d067-4155-901f-e67ea91eeb2b req-c50fe925-b5b0-47e8-af27-735650fffc7a service nova] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Received unexpected event network-vif-plugged-fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8 for instance with vm_state building and task_state spawning. [ 843.851409] env[62692]: INFO nova.compute.manager [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Took 31.33 seconds to build instance. [ 843.978260] env[62692]: DEBUG nova.network.neutron [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Successfully updated port: fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 844.115663] env[62692]: DEBUG nova.compute.utils [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 844.117643] env[62692]: DEBUG nova.compute.manager [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 844.117643] env[62692]: DEBUG nova.network.neutron [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 844.198394] env[62692]: DEBUG nova.policy [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '249a3b566fcf4d88a973693512df370e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0b0b2e09a06a4fc7a00a03f72989ffed', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 844.212124] env[62692]: DEBUG oslo_vmware.api [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141149, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.238334] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Task: {'id': task-1141148, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.817055} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.238625] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 7a3f9c7d-03cf-4177-88db-acf33cc03feb/7a3f9c7d-03cf-4177-88db-acf33cc03feb.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 844.238933] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 844.239281] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0b494039-897f-4417-9901-d2d31f2cac75 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.246702] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Waiting for the task: (returnval){ [ 844.246702] env[62692]: value = "task-1141151" [ 844.246702] env[62692]: _type = "Task" [ 844.246702] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.257280] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Task: {'id': task-1141151, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.353426] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4cdc2c83-4664-4c46-aa10-33cb69b949a3 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "8b75e781-930e-4885-81d2-8a8929d6c39b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.388s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.486153] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Acquiring lock "refresh_cache-c6a965d3-1c52-4f95-a226-9d15b7197ce7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.486153] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Acquired lock "refresh_cache-c6a965d3-1c52-4f95-a226-9d15b7197ce7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.486153] env[62692]: DEBUG nova.network.neutron [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 844.533247] env[62692]: DEBUG nova.network.neutron [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Successfully created port: 99d28285-5d04-45ee-a29f-40f81babef5f {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 844.587682] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "8b75e781-930e-4885-81d2-8a8929d6c39b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.588183] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "8b75e781-930e-4885-81d2-8a8929d6c39b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.588501] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "8b75e781-930e-4885-81d2-8a8929d6c39b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.588811] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "8b75e781-930e-4885-81d2-8a8929d6c39b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.589228] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "8b75e781-930e-4885-81d2-8a8929d6c39b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.592127] env[62692]: INFO nova.compute.manager [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Terminating instance [ 844.595243] env[62692]: DEBUG nova.compute.manager [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 844.595444] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 844.596275] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24ec43e2-0b95-4efb-b6da-025b88438f14 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.604821] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 844.605076] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-146bde4f-bd23-42ae-9137-182345105a38 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.611795] env[62692]: DEBUG oslo_vmware.api [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 844.611795] env[62692]: value = "task-1141152" [ 844.611795] env[62692]: _type = "Task" [ 844.611795] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.621549] env[62692]: DEBUG oslo_vmware.api [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141152, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.623249] env[62692]: DEBUG nova.compute.manager [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 844.710187] env[62692]: DEBUG oslo_vmware.api [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141149, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.784227} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.713056] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 2ccb76cf-f641-4306-a137-fb2417285df9/2ccb76cf-f641-4306-a137-fb2417285df9.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 844.713467] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 844.714156] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d2a0b2c3-b04e-4f12-9864-f2eecc0644d1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.722149] env[62692]: DEBUG oslo_vmware.api [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Waiting for the task: (returnval){ [ 844.722149] env[62692]: value = "task-1141153" [ 844.722149] env[62692]: _type = "Task" [ 844.722149] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.733092] env[62692]: DEBUG oslo_vmware.api [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141153, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.756761] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Task: {'id': task-1141151, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.114244} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.757103] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 844.758019] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4edc1a23-b41e-4813-a770-3ff996532940 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.781232] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Reconfiguring VM instance instance-00000031 to attach disk [datastore2] 7a3f9c7d-03cf-4177-88db-acf33cc03feb/7a3f9c7d-03cf-4177-88db-acf33cc03feb.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 844.784680] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ead95381-3188-4252-b677-d3d819aa0164 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.806596] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Waiting for the task: (returnval){ [ 844.806596] env[62692]: value = "task-1141154" [ 844.806596] env[62692]: _type = "Task" [ 844.806596] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.815367] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Task: {'id': task-1141154, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.857839] env[62692]: DEBUG nova.compute.manager [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 845.031633] env[62692]: DEBUG nova.network.neutron [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 845.063033] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14c96b60-5ea3-4ee0-96b8-297068c09781 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.070903] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-943869e1-c31c-4655-8bb3-9b72df2bcfbb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.103684] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4a7e2b6-96ba-4181-9601-05fecd439a5e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.111084] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd4080a8-99c7-4f5c-85d9-236c6e732445 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.127098] env[62692]: DEBUG nova.compute.provider_tree [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 845.133685] env[62692]: DEBUG oslo_vmware.api [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141152, 'name': PowerOffVM_Task, 'duration_secs': 0.202129} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.136470] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 845.136557] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 845.136856] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-663b0ad7-21ca-4157-8c1d-97a400797e04 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.198562] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 845.198842] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 845.199076] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Deleting the datastore file [datastore1] 8b75e781-930e-4885-81d2-8a8929d6c39b {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 845.199637] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7b7d2bf6-c6eb-45d7-bca6-005c1cf19f92 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.205552] env[62692]: DEBUG oslo_vmware.api [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 845.205552] env[62692]: value = "task-1141156" [ 845.205552] env[62692]: _type = "Task" [ 845.205552] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.213210] env[62692]: DEBUG oslo_vmware.api [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141156, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.216960] env[62692]: DEBUG nova.network.neutron [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Updating instance_info_cache with network_info: [{"id": "fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8", "address": "fa:16:3e:c7:7a:a0", "network": {"id": "04aacc3d-37b6-4a5c-ac4d-55313731e4f9", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1367536055-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef45ad1482dc4f5db51a5badc3cdce22", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a0734cc4-5718-45e2-9f98-0ded96880bef", "external-id": "nsx-vlan-transportzone-875", "segmentation_id": 875, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdd5ae79-fe", "ovs_interfaceid": "fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.230387] env[62692]: DEBUG oslo_vmware.api [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141153, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079839} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.230650] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 845.231491] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49e65d78-dfcc-4db9-a060-06ce5ae57c85 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.254374] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] 2ccb76cf-f641-4306-a137-fb2417285df9/2ccb76cf-f641-4306-a137-fb2417285df9.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 845.254826] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-458a5994-79f0-449d-9751-931c3e178efd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.274415] env[62692]: DEBUG oslo_vmware.api [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Waiting for the task: (returnval){ [ 845.274415] env[62692]: value = "task-1141157" [ 845.274415] env[62692]: _type = "Task" [ 845.274415] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.283426] env[62692]: DEBUG oslo_vmware.api [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141157, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.316156] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Task: {'id': task-1141154, 'name': ReconfigVM_Task, 'duration_secs': 0.306889} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.316497] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Reconfigured VM instance instance-00000031 to attach disk [datastore2] 7a3f9c7d-03cf-4177-88db-acf33cc03feb/7a3f9c7d-03cf-4177-88db-acf33cc03feb.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 845.317187] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f7384134-cb3c-4780-8f3f-083bcac14b0d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.323263] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Waiting for the task: (returnval){ [ 845.323263] env[62692]: value = "task-1141158" [ 845.323263] env[62692]: _type = "Task" [ 845.323263] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.331436] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Task: {'id': task-1141158, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.382365] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.636127] env[62692]: DEBUG nova.scheduler.client.report [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 845.640306] env[62692]: DEBUG nova.compute.manager [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 845.667782] env[62692]: DEBUG nova.virt.hardware [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 845.668099] env[62692]: DEBUG nova.virt.hardware [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 845.668269] env[62692]: DEBUG nova.virt.hardware [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 845.668459] env[62692]: DEBUG nova.virt.hardware [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 845.668606] env[62692]: DEBUG nova.virt.hardware [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 845.668818] env[62692]: DEBUG nova.virt.hardware [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 845.669268] env[62692]: DEBUG nova.virt.hardware [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 845.669453] env[62692]: DEBUG nova.virt.hardware [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 845.669627] env[62692]: DEBUG nova.virt.hardware [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 845.669824] env[62692]: DEBUG nova.virt.hardware [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 845.670040] env[62692]: DEBUG nova.virt.hardware [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 845.670904] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9df21c9a-84cf-4c35-86da-d1a07e11dd85 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.679290] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dfb4945-0a06-4a26-8e9d-6eb07c0eec23 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.714926] env[62692]: DEBUG oslo_vmware.api [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141156, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.719658] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Releasing lock "refresh_cache-c6a965d3-1c52-4f95-a226-9d15b7197ce7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.719898] env[62692]: DEBUG nova.compute.manager [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Instance network_info: |[{"id": "fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8", "address": "fa:16:3e:c7:7a:a0", "network": {"id": "04aacc3d-37b6-4a5c-ac4d-55313731e4f9", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1367536055-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef45ad1482dc4f5db51a5badc3cdce22", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a0734cc4-5718-45e2-9f98-0ded96880bef", "external-id": "nsx-vlan-transportzone-875", "segmentation_id": 875, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdd5ae79-fe", "ovs_interfaceid": "fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 845.720306] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c7:7a:a0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a0734cc4-5718-45e2-9f98-0ded96880bef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 845.728552] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Creating folder: Project (ef45ad1482dc4f5db51a5badc3cdce22). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 845.728552] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-22c41e42-139d-4898-b90f-b3a80b5f33ce {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.738448] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Created folder: Project (ef45ad1482dc4f5db51a5badc3cdce22) in parent group-v248868. [ 845.738626] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Creating folder: Instances. Parent ref: group-v248944. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 845.738881] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-095f2af4-d0c8-4d8b-a1f7-9befc8b4dd7f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.747810] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Created folder: Instances in parent group-v248944. [ 845.747810] env[62692]: DEBUG oslo.service.loopingcall [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 845.747810] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 845.747979] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6f0054de-c6d1-47be-a738-8cd28b85a21d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.766556] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 845.766556] env[62692]: value = "task-1141161" [ 845.766556] env[62692]: _type = "Task" [ 845.766556] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.776339] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141161, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.783649] env[62692]: DEBUG oslo_vmware.api [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141157, 'name': ReconfigVM_Task, 'duration_secs': 0.381979} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.783915] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Reconfigured VM instance instance-00000035 to attach disk [datastore1] 2ccb76cf-f641-4306-a137-fb2417285df9/2ccb76cf-f641-4306-a137-fb2417285df9.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 845.785021] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4f14c9ec-3e28-4df0-8ac4-46ff40472271 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.796334] env[62692]: DEBUG oslo_vmware.api [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Waiting for the task: (returnval){ [ 845.796334] env[62692]: value = "task-1141162" [ 845.796334] env[62692]: _type = "Task" [ 845.796334] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.804156] env[62692]: DEBUG oslo_vmware.api [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141162, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.833153] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Task: {'id': task-1141158, 'name': Rename_Task, 'duration_secs': 0.1371} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.833451] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 845.833696] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-32216116-56c8-4d03-a14d-1cbdc9f505d6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.842046] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Waiting for the task: (returnval){ [ 845.842046] env[62692]: value = "task-1141163" [ 845.842046] env[62692]: _type = "Task" [ 845.842046] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.850212] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Task: {'id': task-1141163, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.893491] env[62692]: DEBUG nova.compute.manager [req-a660eaa2-e765-4ed5-9f2b-466ab47b641d req-0b1da43e-af24-4f05-acc7-b294e84ec2a9 service nova] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Received event network-changed-fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 845.893696] env[62692]: DEBUG nova.compute.manager [req-a660eaa2-e765-4ed5-9f2b-466ab47b641d req-0b1da43e-af24-4f05-acc7-b294e84ec2a9 service nova] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Refreshing instance network info cache due to event network-changed-fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 845.893928] env[62692]: DEBUG oslo_concurrency.lockutils [req-a660eaa2-e765-4ed5-9f2b-466ab47b641d req-0b1da43e-af24-4f05-acc7-b294e84ec2a9 service nova] Acquiring lock "refresh_cache-c6a965d3-1c52-4f95-a226-9d15b7197ce7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.894106] env[62692]: DEBUG oslo_concurrency.lockutils [req-a660eaa2-e765-4ed5-9f2b-466ab47b641d req-0b1da43e-af24-4f05-acc7-b294e84ec2a9 service nova] Acquired lock "refresh_cache-c6a965d3-1c52-4f95-a226-9d15b7197ce7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.894275] env[62692]: DEBUG nova.network.neutron [req-a660eaa2-e765-4ed5-9f2b-466ab47b641d req-0b1da43e-af24-4f05-acc7-b294e84ec2a9 service nova] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Refreshing network info cache for port fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 846.145297] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.536s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.145582] env[62692]: DEBUG nova.compute.manager [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 846.153293] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.054s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.154932] env[62692]: INFO nova.compute.claims [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 846.223027] env[62692]: DEBUG oslo_vmware.api [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141156, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.634814} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.223027] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 846.223027] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 846.223027] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 846.223027] env[62692]: INFO nova.compute.manager [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Took 1.63 seconds to destroy the instance on the hypervisor. [ 846.223027] env[62692]: DEBUG oslo.service.loopingcall [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 846.223027] env[62692]: DEBUG nova.compute.manager [-] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 846.223027] env[62692]: DEBUG nova.network.neutron [-] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 846.241606] env[62692]: DEBUG nova.network.neutron [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Successfully updated port: 99d28285-5d04-45ee-a29f-40f81babef5f {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 846.279135] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141161, 'name': CreateVM_Task, 'duration_secs': 0.350824} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.279135] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 846.279135] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.279466] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.279619] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 846.280194] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66a507a7-dab8-4f01-8f97-29d45e8614e6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.286309] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Waiting for the task: (returnval){ [ 846.286309] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52b09274-0103-ed6a-292e-dc1e317f1603" [ 846.286309] env[62692]: _type = "Task" [ 846.286309] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.295269] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52b09274-0103-ed6a-292e-dc1e317f1603, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.301455] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-010bcb5f-4168-4c3d-88d0-9f094c7e6b68 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Volume attach. Driver type: vmdk {{(pid=62692) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 846.301771] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-010bcb5f-4168-4c3d-88d0-9f094c7e6b68 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-248943', 'volume_id': 'eccf2e07-c766-49a3-9929-1c9805b08d2a', 'name': 'volume-eccf2e07-c766-49a3-9929-1c9805b08d2a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'abee88ff-5cf5-4bf1-91e0-93b19cf30046', 'attached_at': '', 'detached_at': '', 'volume_id': 'eccf2e07-c766-49a3-9929-1c9805b08d2a', 'serial': 'eccf2e07-c766-49a3-9929-1c9805b08d2a'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 846.302629] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c69179-19e5-4758-9572-4b9ade0d705e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.310410] env[62692]: DEBUG oslo_vmware.api [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141162, 'name': Rename_Task, 'duration_secs': 0.201917} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.320782] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 846.321233] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a16b505d-12de-4f7e-adbe-ed1f5c7d28fd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.323456] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc539ca6-75ce-4482-9e3e-a033ffd1323c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.351951] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-010bcb5f-4168-4c3d-88d0-9f094c7e6b68 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Reconfiguring VM instance instance-00000023 to attach disk [datastore2] volume-eccf2e07-c766-49a3-9929-1c9805b08d2a/volume-eccf2e07-c766-49a3-9929-1c9805b08d2a.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 846.353566] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ce7a2a34-7610-4bbb-99e4-a2a2357ef114 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.366983] env[62692]: DEBUG oslo_vmware.api [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Waiting for the task: (returnval){ [ 846.366983] env[62692]: value = "task-1141164" [ 846.366983] env[62692]: _type = "Task" [ 846.366983] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.376665] env[62692]: DEBUG oslo_vmware.api [None req-010bcb5f-4168-4c3d-88d0-9f094c7e6b68 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 846.376665] env[62692]: value = "task-1141165" [ 846.376665] env[62692]: _type = "Task" [ 846.376665] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.382213] env[62692]: DEBUG oslo_vmware.api [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141164, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.382213] env[62692]: DEBUG oslo_vmware.api [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Task: {'id': task-1141163, 'name': PowerOnVM_Task, 'duration_secs': 0.469657} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.384713] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 846.384931] env[62692]: DEBUG nova.compute.manager [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 846.385655] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1643576b-7ba3-4ec8-9ef1-f49bcb2b67ac {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.394733] env[62692]: DEBUG oslo_vmware.api [None req-010bcb5f-4168-4c3d-88d0-9f094c7e6b68 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1141165, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.550560] env[62692]: DEBUG nova.compute.manager [req-a314d012-8a4d-400b-96f5-2ea6b7b18dc4 req-941cc072-45d4-443b-8d82-94fe0976f336 service nova] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Received event network-vif-deleted-4eb54953-e726-4bd7-9793-30a73f994d30 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 846.550829] env[62692]: INFO nova.compute.manager [req-a314d012-8a4d-400b-96f5-2ea6b7b18dc4 req-941cc072-45d4-443b-8d82-94fe0976f336 service nova] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Neutron deleted interface 4eb54953-e726-4bd7-9793-30a73f994d30; detaching it from the instance and deleting it from the info cache [ 846.550959] env[62692]: DEBUG nova.network.neutron [req-a314d012-8a4d-400b-96f5-2ea6b7b18dc4 req-941cc072-45d4-443b-8d82-94fe0976f336 service nova] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.663710] env[62692]: DEBUG nova.compute.utils [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 846.665466] env[62692]: DEBUG nova.compute.manager [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 846.665640] env[62692]: DEBUG nova.network.neutron [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 846.680256] env[62692]: DEBUG nova.network.neutron [req-a660eaa2-e765-4ed5-9f2b-466ab47b641d req-0b1da43e-af24-4f05-acc7-b294e84ec2a9 service nova] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Updated VIF entry in instance network info cache for port fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 846.680618] env[62692]: DEBUG nova.network.neutron [req-a660eaa2-e765-4ed5-9f2b-466ab47b641d req-0b1da43e-af24-4f05-acc7-b294e84ec2a9 service nova] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Updating instance_info_cache with network_info: [{"id": "fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8", "address": "fa:16:3e:c7:7a:a0", "network": {"id": "04aacc3d-37b6-4a5c-ac4d-55313731e4f9", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1367536055-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef45ad1482dc4f5db51a5badc3cdce22", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a0734cc4-5718-45e2-9f98-0ded96880bef", "external-id": "nsx-vlan-transportzone-875", "segmentation_id": 875, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdd5ae79-fe", "ovs_interfaceid": "fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.707223] env[62692]: DEBUG nova.policy [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd07bd166be464063b6cf3c39ef0d5e3e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fa9699bac6e44e6c8132c2eb63aca648', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 846.743849] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "refresh_cache-64a8329d-e0e9-4c2d-bd1f-844ee40eb980" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.743977] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquired lock "refresh_cache-64a8329d-e0e9-4c2d-bd1f-844ee40eb980" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.744174] env[62692]: DEBUG nova.network.neutron [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 846.804132] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52b09274-0103-ed6a-292e-dc1e317f1603, 'name': SearchDatastore_Task, 'duration_secs': 0.018793} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.804557] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.804836] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 846.805121] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.805290] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.805486] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 846.805766] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7b1b1c12-27b2-4bbe-9f3f-2089b0cd6ae0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.815428] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 846.815628] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 846.816508] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d630f897-01ed-4257-be3e-27b91666a650 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.822466] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Waiting for the task: (returnval){ [ 846.822466] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52d37eda-018c-65d3-a2b2-87b38c6c28a8" [ 846.822466] env[62692]: _type = "Task" [ 846.822466] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.834022] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52d37eda-018c-65d3-a2b2-87b38c6c28a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.880213] env[62692]: DEBUG oslo_vmware.api [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141164, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.890232] env[62692]: DEBUG oslo_vmware.api [None req-010bcb5f-4168-4c3d-88d0-9f094c7e6b68 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1141165, 'name': ReconfigVM_Task, 'duration_secs': 0.463193} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.890792] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-010bcb5f-4168-4c3d-88d0-9f094c7e6b68 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Reconfigured VM instance instance-00000023 to attach disk [datastore2] volume-eccf2e07-c766-49a3-9929-1c9805b08d2a/volume-eccf2e07-c766-49a3-9929-1c9805b08d2a.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 846.895650] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ba1c41ef-a502-4b19-9743-7f0d81f34619 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.917498] env[62692]: DEBUG oslo_vmware.api [None req-010bcb5f-4168-4c3d-88d0-9f094c7e6b68 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 846.917498] env[62692]: value = "task-1141166" [ 846.917498] env[62692]: _type = "Task" [ 846.917498] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.917981] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.928954] env[62692]: DEBUG oslo_vmware.api [None req-010bcb5f-4168-4c3d-88d0-9f094c7e6b68 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1141166, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.024663] env[62692]: DEBUG nova.network.neutron [-] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.043775] env[62692]: DEBUG nova.network.neutron [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Successfully created port: ddc1a77b-94d9-468b-ab57-8aef5343e580 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 847.053901] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-871b94b2-45d5-473c-84d2-f5ab4ae43db6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.062899] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a517276-d11a-4f20-8f24-f87d036f80e3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.103172] env[62692]: DEBUG nova.compute.manager [req-a314d012-8a4d-400b-96f5-2ea6b7b18dc4 req-941cc072-45d4-443b-8d82-94fe0976f336 service nova] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Detach interface failed, port_id=4eb54953-e726-4bd7-9793-30a73f994d30, reason: Instance 8b75e781-930e-4885-81d2-8a8929d6c39b could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 847.170102] env[62692]: DEBUG nova.compute.manager [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 847.183497] env[62692]: DEBUG oslo_concurrency.lockutils [req-a660eaa2-e765-4ed5-9f2b-466ab47b641d req-0b1da43e-af24-4f05-acc7-b294e84ec2a9 service nova] Releasing lock "refresh_cache-c6a965d3-1c52-4f95-a226-9d15b7197ce7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.295824] env[62692]: DEBUG nova.network.neutron [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 847.333489] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52d37eda-018c-65d3-a2b2-87b38c6c28a8, 'name': SearchDatastore_Task, 'duration_secs': 0.009342} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.337930] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15aadfe5-141a-4aa6-87fc-2f294e6cf2f8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.343525] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Waiting for the task: (returnval){ [ 847.343525] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52c4958a-4730-eb3b-032b-25c1e8450184" [ 847.343525] env[62692]: _type = "Task" [ 847.343525] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.353623] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52c4958a-4730-eb3b-032b-25c1e8450184, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.387165] env[62692]: DEBUG oslo_vmware.api [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141164, 'name': PowerOnVM_Task} progress is 91%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.429533] env[62692]: DEBUG oslo_vmware.api [None req-010bcb5f-4168-4c3d-88d0-9f094c7e6b68 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1141166, 'name': ReconfigVM_Task, 'duration_secs': 0.150756} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.432803] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-010bcb5f-4168-4c3d-88d0-9f094c7e6b68 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-248943', 'volume_id': 'eccf2e07-c766-49a3-9929-1c9805b08d2a', 'name': 'volume-eccf2e07-c766-49a3-9929-1c9805b08d2a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'abee88ff-5cf5-4bf1-91e0-93b19cf30046', 'attached_at': '', 'detached_at': '', 'volume_id': 'eccf2e07-c766-49a3-9929-1c9805b08d2a', 'serial': 'eccf2e07-c766-49a3-9929-1c9805b08d2a'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 847.526150] env[62692]: DEBUG nova.network.neutron [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Updating instance_info_cache with network_info: [{"id": "99d28285-5d04-45ee-a29f-40f81babef5f", "address": "fa:16:3e:bc:9b:28", "network": {"id": "2c82845f-101b-48c4-bd0b-4704573a3615", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1972972124-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b0b2e09a06a4fc7a00a03f72989ffed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16f09e8c-5240-4839-80cc-62ec29700bd2", "external-id": "nsx-vlan-transportzone-720", "segmentation_id": 720, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99d28285-5d", "ovs_interfaceid": "99d28285-5d04-45ee-a29f-40f81babef5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.527773] env[62692]: INFO nova.compute.manager [-] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Took 1.30 seconds to deallocate network for instance. [ 847.635606] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65d53568-78e7-4b30-8227-7e2aaea3778e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.643974] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c3de95e-7637-4abd-ab50-df27831c7457 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.676784] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7c6fa7f-0c7d-4d27-a98c-687d4b31c3e6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.688389] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-578d85ed-b83c-4b61-8ad4-42b54d6b02ec {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.701849] env[62692]: DEBUG nova.compute.provider_tree [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 847.854910] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52c4958a-4730-eb3b-032b-25c1e8450184, 'name': SearchDatastore_Task, 'duration_secs': 0.025318} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.855204] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.855493] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] c6a965d3-1c52-4f95-a226-9d15b7197ce7/c6a965d3-1c52-4f95-a226-9d15b7197ce7.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 847.855743] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fb6bec16-392f-4348-bf46-6b51b3221df2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.860373] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Acquiring lock "7a3f9c7d-03cf-4177-88db-acf33cc03feb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.860603] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Lock "7a3f9c7d-03cf-4177-88db-acf33cc03feb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.860826] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Acquiring lock "7a3f9c7d-03cf-4177-88db-acf33cc03feb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.860999] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Lock "7a3f9c7d-03cf-4177-88db-acf33cc03feb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.861179] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Lock "7a3f9c7d-03cf-4177-88db-acf33cc03feb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.863294] env[62692]: INFO nova.compute.manager [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Terminating instance [ 847.866472] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Waiting for the task: (returnval){ [ 847.866472] env[62692]: value = "task-1141167" [ 847.866472] env[62692]: _type = "Task" [ 847.866472] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.866910] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Acquiring lock "refresh_cache-7a3f9c7d-03cf-4177-88db-acf33cc03feb" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.867160] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Acquired lock "refresh_cache-7a3f9c7d-03cf-4177-88db-acf33cc03feb" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.867347] env[62692]: DEBUG nova.network.neutron [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 847.878064] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Task: {'id': task-1141167, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.883201] env[62692]: DEBUG oslo_vmware.api [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141164, 'name': PowerOnVM_Task, 'duration_secs': 1.051092} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.883506] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 847.883664] env[62692]: INFO nova.compute.manager [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Took 9.45 seconds to spawn the instance on the hypervisor. [ 847.883867] env[62692]: DEBUG nova.compute.manager [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 847.884617] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05bb6e5f-beee-4072-9b4c-dd8af0a8fbec {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.928655] env[62692]: DEBUG nova.compute.manager [req-284f913d-777c-4a86-8283-fd9968f931f8 req-4833d0e5-ebfc-4813-a140-99435a607dc1 service nova] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Received event network-vif-plugged-99d28285-5d04-45ee-a29f-40f81babef5f {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 847.928962] env[62692]: DEBUG oslo_concurrency.lockutils [req-284f913d-777c-4a86-8283-fd9968f931f8 req-4833d0e5-ebfc-4813-a140-99435a607dc1 service nova] Acquiring lock "64a8329d-e0e9-4c2d-bd1f-844ee40eb980-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.929195] env[62692]: DEBUG oslo_concurrency.lockutils [req-284f913d-777c-4a86-8283-fd9968f931f8 req-4833d0e5-ebfc-4813-a140-99435a607dc1 service nova] Lock "64a8329d-e0e9-4c2d-bd1f-844ee40eb980-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.929381] env[62692]: DEBUG oslo_concurrency.lockutils [req-284f913d-777c-4a86-8283-fd9968f931f8 req-4833d0e5-ebfc-4813-a140-99435a607dc1 service nova] Lock "64a8329d-e0e9-4c2d-bd1f-844ee40eb980-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.929517] env[62692]: DEBUG nova.compute.manager [req-284f913d-777c-4a86-8283-fd9968f931f8 req-4833d0e5-ebfc-4813-a140-99435a607dc1 service nova] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] No waiting events found dispatching network-vif-plugged-99d28285-5d04-45ee-a29f-40f81babef5f {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 847.929676] env[62692]: WARNING nova.compute.manager [req-284f913d-777c-4a86-8283-fd9968f931f8 req-4833d0e5-ebfc-4813-a140-99435a607dc1 service nova] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Received unexpected event network-vif-plugged-99d28285-5d04-45ee-a29f-40f81babef5f for instance with vm_state building and task_state spawning. [ 847.929827] env[62692]: DEBUG nova.compute.manager [req-284f913d-777c-4a86-8283-fd9968f931f8 req-4833d0e5-ebfc-4813-a140-99435a607dc1 service nova] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Received event network-changed-99d28285-5d04-45ee-a29f-40f81babef5f {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 847.929979] env[62692]: DEBUG nova.compute.manager [req-284f913d-777c-4a86-8283-fd9968f931f8 req-4833d0e5-ebfc-4813-a140-99435a607dc1 service nova] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Refreshing instance network info cache due to event network-changed-99d28285-5d04-45ee-a29f-40f81babef5f. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 847.930177] env[62692]: DEBUG oslo_concurrency.lockutils [req-284f913d-777c-4a86-8283-fd9968f931f8 req-4833d0e5-ebfc-4813-a140-99435a607dc1 service nova] Acquiring lock "refresh_cache-64a8329d-e0e9-4c2d-bd1f-844ee40eb980" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.029437] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Releasing lock "refresh_cache-64a8329d-e0e9-4c2d-bd1f-844ee40eb980" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.029437] env[62692]: DEBUG nova.compute.manager [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Instance network_info: |[{"id": "99d28285-5d04-45ee-a29f-40f81babef5f", "address": "fa:16:3e:bc:9b:28", "network": {"id": "2c82845f-101b-48c4-bd0b-4704573a3615", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1972972124-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b0b2e09a06a4fc7a00a03f72989ffed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16f09e8c-5240-4839-80cc-62ec29700bd2", "external-id": "nsx-vlan-transportzone-720", "segmentation_id": 720, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99d28285-5d", "ovs_interfaceid": "99d28285-5d04-45ee-a29f-40f81babef5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 848.029437] env[62692]: DEBUG oslo_concurrency.lockutils [req-284f913d-777c-4a86-8283-fd9968f931f8 req-4833d0e5-ebfc-4813-a140-99435a607dc1 service nova] Acquired lock "refresh_cache-64a8329d-e0e9-4c2d-bd1f-844ee40eb980" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.029437] env[62692]: DEBUG nova.network.neutron [req-284f913d-777c-4a86-8283-fd9968f931f8 req-4833d0e5-ebfc-4813-a140-99435a607dc1 service nova] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Refreshing network info cache for port 99d28285-5d04-45ee-a29f-40f81babef5f {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 848.029802] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bc:9b:28', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '16f09e8c-5240-4839-80cc-62ec29700bd2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '99d28285-5d04-45ee-a29f-40f81babef5f', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 848.037169] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Creating folder: Project (0b0b2e09a06a4fc7a00a03f72989ffed). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 848.041138] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.041400] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-71d5cd22-8c19-4bf6-9aef-bc9edf7ae238 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.052565] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Created folder: Project (0b0b2e09a06a4fc7a00a03f72989ffed) in parent group-v248868. [ 848.052792] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Creating folder: Instances. Parent ref: group-v248947. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 848.053016] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-db5e3e6b-4a93-4623-ae53-f28f206090c1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.062563] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Created folder: Instances in parent group-v248947. [ 848.062806] env[62692]: DEBUG oslo.service.loopingcall [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 848.063013] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 848.063510] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-35e5d024-dc84-4cde-85da-c53399ef5d03 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.083730] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 848.083730] env[62692]: value = "task-1141170" [ 848.083730] env[62692]: _type = "Task" [ 848.083730] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.092855] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141170, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.184672] env[62692]: DEBUG nova.compute.manager [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 848.204833] env[62692]: DEBUG nova.scheduler.client.report [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 848.222813] env[62692]: DEBUG nova.virt.hardware [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 848.223093] env[62692]: DEBUG nova.virt.hardware [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 848.223270] env[62692]: DEBUG nova.virt.hardware [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 848.223461] env[62692]: DEBUG nova.virt.hardware [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 848.223660] env[62692]: DEBUG nova.virt.hardware [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 848.224031] env[62692]: DEBUG nova.virt.hardware [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 848.224617] env[62692]: DEBUG nova.virt.hardware [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 848.224893] env[62692]: DEBUG nova.virt.hardware [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 848.225186] env[62692]: DEBUG nova.virt.hardware [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 848.225469] env[62692]: DEBUG nova.virt.hardware [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 848.225775] env[62692]: DEBUG nova.virt.hardware [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 848.227102] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67baa158-9181-4207-a374-703b6606a0ed {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.239065] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0da6a9a8-1aaf-4b40-9b8c-330c67e2215d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.383618] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Task: {'id': task-1141167, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.399734] env[62692]: DEBUG nova.network.neutron [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 848.404198] env[62692]: INFO nova.compute.manager [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Took 34.58 seconds to build instance. [ 848.478541] env[62692]: DEBUG nova.objects.instance [None req-010bcb5f-4168-4c3d-88d0-9f094c7e6b68 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lazy-loading 'flavor' on Instance uuid abee88ff-5cf5-4bf1-91e0-93b19cf30046 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 848.491679] env[62692]: DEBUG nova.network.neutron [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.497835] env[62692]: DEBUG nova.network.neutron [req-284f913d-777c-4a86-8283-fd9968f931f8 req-4833d0e5-ebfc-4813-a140-99435a607dc1 service nova] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Updated VIF entry in instance network info cache for port 99d28285-5d04-45ee-a29f-40f81babef5f. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 848.498135] env[62692]: DEBUG nova.network.neutron [req-284f913d-777c-4a86-8283-fd9968f931f8 req-4833d0e5-ebfc-4813-a140-99435a607dc1 service nova] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Updating instance_info_cache with network_info: [{"id": "99d28285-5d04-45ee-a29f-40f81babef5f", "address": "fa:16:3e:bc:9b:28", "network": {"id": "2c82845f-101b-48c4-bd0b-4704573a3615", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1972972124-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b0b2e09a06a4fc7a00a03f72989ffed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16f09e8c-5240-4839-80cc-62ec29700bd2", "external-id": "nsx-vlan-transportzone-720", "segmentation_id": 720, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99d28285-5d", "ovs_interfaceid": "99d28285-5d04-45ee-a29f-40f81babef5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.596234] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141170, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.633546] env[62692]: DEBUG nova.compute.manager [req-691b23b2-d401-4289-9911-a373de969626 req-13ee1eb2-14bd-4a87-adbd-3a7a1c3a1f4b service nova] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Received event network-vif-plugged-ddc1a77b-94d9-468b-ab57-8aef5343e580 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 848.633546] env[62692]: DEBUG oslo_concurrency.lockutils [req-691b23b2-d401-4289-9911-a373de969626 req-13ee1eb2-14bd-4a87-adbd-3a7a1c3a1f4b service nova] Acquiring lock "18eb004f-a16b-477a-8d05-ed23a418b069-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.633546] env[62692]: DEBUG oslo_concurrency.lockutils [req-691b23b2-d401-4289-9911-a373de969626 req-13ee1eb2-14bd-4a87-adbd-3a7a1c3a1f4b service nova] Lock "18eb004f-a16b-477a-8d05-ed23a418b069-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.633546] env[62692]: DEBUG oslo_concurrency.lockutils [req-691b23b2-d401-4289-9911-a373de969626 req-13ee1eb2-14bd-4a87-adbd-3a7a1c3a1f4b service nova] Lock "18eb004f-a16b-477a-8d05-ed23a418b069-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.633546] env[62692]: DEBUG nova.compute.manager [req-691b23b2-d401-4289-9911-a373de969626 req-13ee1eb2-14bd-4a87-adbd-3a7a1c3a1f4b service nova] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] No waiting events found dispatching network-vif-plugged-ddc1a77b-94d9-468b-ab57-8aef5343e580 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 848.633546] env[62692]: WARNING nova.compute.manager [req-691b23b2-d401-4289-9911-a373de969626 req-13ee1eb2-14bd-4a87-adbd-3a7a1c3a1f4b service nova] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Received unexpected event network-vif-plugged-ddc1a77b-94d9-468b-ab57-8aef5343e580 for instance with vm_state building and task_state spawning. [ 848.711782] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.558s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.713028] env[62692]: DEBUG nova.compute.manager [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 848.718098] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.944s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.718098] env[62692]: INFO nova.compute.claims [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 848.743896] env[62692]: DEBUG nova.network.neutron [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Successfully updated port: ddc1a77b-94d9-468b-ab57-8aef5343e580 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 848.882414] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Task: {'id': task-1141167, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.669656} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.882620] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] c6a965d3-1c52-4f95-a226-9d15b7197ce7/c6a965d3-1c52-4f95-a226-9d15b7197ce7.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 848.882859] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 848.883138] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-03d15da3-2e8b-42bc-ba8e-53f5b7f0ee38 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.893188] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Waiting for the task: (returnval){ [ 848.893188] env[62692]: value = "task-1141171" [ 848.893188] env[62692]: _type = "Task" [ 848.893188] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.901941] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Task: {'id': task-1141171, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.906536] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3dd8c678-d4ab-4d20-9911-d828ce89b212 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Lock "2ccb76cf-f641-4306-a137-fb2417285df9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.218s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.983752] env[62692]: DEBUG oslo_concurrency.lockutils [None req-010bcb5f-4168-4c3d-88d0-9f094c7e6b68 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lock "abee88ff-5cf5-4bf1-91e0-93b19cf30046" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.314s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.995117] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Releasing lock "refresh_cache-7a3f9c7d-03cf-4177-88db-acf33cc03feb" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.995117] env[62692]: DEBUG nova.compute.manager [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 848.995117] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 848.995699] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b43be15a-a225-40df-8863-3a5e908684dc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.002479] env[62692]: DEBUG oslo_concurrency.lockutils [req-284f913d-777c-4a86-8283-fd9968f931f8 req-4833d0e5-ebfc-4813-a140-99435a607dc1 service nova] Releasing lock "refresh_cache-64a8329d-e0e9-4c2d-bd1f-844ee40eb980" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.006997] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 849.006997] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ec49000f-2204-4f99-b764-9f1344b55278 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.013024] env[62692]: DEBUG oslo_vmware.api [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Waiting for the task: (returnval){ [ 849.013024] env[62692]: value = "task-1141172" [ 849.013024] env[62692]: _type = "Task" [ 849.013024] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.021941] env[62692]: DEBUG oslo_vmware.api [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141172, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.094269] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141170, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.223397] env[62692]: DEBUG nova.compute.utils [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 849.231202] env[62692]: DEBUG nova.compute.manager [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 849.231939] env[62692]: DEBUG nova.network.neutron [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 849.245499] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Acquiring lock "refresh_cache-18eb004f-a16b-477a-8d05-ed23a418b069" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.245499] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Acquired lock "refresh_cache-18eb004f-a16b-477a-8d05-ed23a418b069" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.245499] env[62692]: DEBUG nova.network.neutron [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 849.289681] env[62692]: DEBUG nova.policy [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '016134ac36fd4bd893226ae95a955e28', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37921f67664544478cb2568def068b11', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 849.405708] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Task: {'id': task-1141171, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.409269] env[62692]: DEBUG nova.compute.manager [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 849.523965] env[62692]: DEBUG oslo_vmware.api [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141172, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.594296] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141170, 'name': CreateVM_Task, 'duration_secs': 1.421519} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.594514] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 849.595270] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.595494] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.595947] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 849.596287] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1022094-af09-4e8b-ac8d-ec0397591801 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.602296] env[62692]: DEBUG oslo_vmware.api [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 849.602296] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52ec3c63-a917-fc01-0d12-2cf75fc4a573" [ 849.602296] env[62692]: _type = "Task" [ 849.602296] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.609187] env[62692]: DEBUG oslo_vmware.api [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52ec3c63-a917-fc01-0d12-2cf75fc4a573, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.710561] env[62692]: DEBUG nova.network.neutron [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Successfully created port: a291ddc4-8996-4755-b347-066f38e6ec7d {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 849.731962] env[62692]: DEBUG nova.compute.manager [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 849.735418] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f8208b7e-450d-40e9-903e-37b80dc6f3be tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Acquiring lock "abee88ff-5cf5-4bf1-91e0-93b19cf30046" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.735656] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f8208b7e-450d-40e9-903e-37b80dc6f3be tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lock "abee88ff-5cf5-4bf1-91e0-93b19cf30046" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.796803] env[62692]: DEBUG nova.network.neutron [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 849.907974] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Task: {'id': task-1141171, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.660243} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.908496] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 849.909537] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c8dd0d0-dd42-4e10-9f88-4423ed316f77 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.941845] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Reconfiguring VM instance instance-00000036 to attach disk [datastore2] c6a965d3-1c52-4f95-a226-9d15b7197ce7/c6a965d3-1c52-4f95-a226-9d15b7197ce7.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 849.945557] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.945819] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd527471-b550-4ad2-afc3-17d1e8c9881f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.968171] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Waiting for the task: (returnval){ [ 849.968171] env[62692]: value = "task-1141173" [ 849.968171] env[62692]: _type = "Task" [ 849.968171] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.977427] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Task: {'id': task-1141173, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.025642] env[62692]: DEBUG oslo_vmware.api [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141172, 'name': PowerOffVM_Task, 'duration_secs': 0.578192} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.030198] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 850.030198] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 850.030198] env[62692]: DEBUG nova.network.neutron [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Updating instance_info_cache with network_info: [{"id": "ddc1a77b-94d9-468b-ab57-8aef5343e580", "address": "fa:16:3e:4a:ca:b3", "network": {"id": "8628e5f2-19ef-4119-9620-67957a0fe952", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-559068606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa9699bac6e44e6c8132c2eb63aca648", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddc1a77b-94", "ovs_interfaceid": "ddc1a77b-94d9-468b-ab57-8aef5343e580", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.031679] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b04681c8-f2ea-4ea0-ba7b-7c868430a153 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.070751] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 850.070751] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 850.070751] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Deleting the datastore file [datastore2] 7a3f9c7d-03cf-4177-88db-acf33cc03feb {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 850.071137] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a5fa63fa-32df-468e-a268-817d16e3e44a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.078467] env[62692]: DEBUG oslo_vmware.api [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Waiting for the task: (returnval){ [ 850.078467] env[62692]: value = "task-1141175" [ 850.078467] env[62692]: _type = "Task" [ 850.078467] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.088132] env[62692]: DEBUG oslo_vmware.api [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141175, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.119039] env[62692]: DEBUG oslo_vmware.api [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52ec3c63-a917-fc01-0d12-2cf75fc4a573, 'name': SearchDatastore_Task, 'duration_secs': 0.015208} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.119039] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.119039] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 850.119039] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.119039] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.119039] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 850.119039] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cbeeeda7-b28c-41fc-8a0b-d1827d8b7f80 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.136874] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 850.137635] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 850.137906] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc356967-7a33-4550-a25c-cef8eefa0304 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.143951] env[62692]: DEBUG oslo_vmware.api [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 850.143951] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]522f7b2d-4c01-5c6d-7a7b-e260560a24af" [ 850.143951] env[62692]: _type = "Task" [ 850.143951] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.155924] env[62692]: DEBUG oslo_vmware.api [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]522f7b2d-4c01-5c6d-7a7b-e260560a24af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.241357] env[62692]: INFO nova.compute.manager [None req-f8208b7e-450d-40e9-903e-37b80dc6f3be tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Detaching volume eccf2e07-c766-49a3-9929-1c9805b08d2a [ 850.273583] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-024ee3a6-6b0d-4691-92ce-f2be573a6ac0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.285045] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-189cd473-1b9b-4e68-a105-177a08b4bed6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.320551] env[62692]: INFO nova.virt.block_device [None req-f8208b7e-450d-40e9-903e-37b80dc6f3be tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Attempting to driver detach volume eccf2e07-c766-49a3-9929-1c9805b08d2a from mountpoint /dev/sdb [ 850.320804] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8208b7e-450d-40e9-903e-37b80dc6f3be tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Volume detach. Driver type: vmdk {{(pid=62692) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 850.320981] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8208b7e-450d-40e9-903e-37b80dc6f3be tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-248943', 'volume_id': 'eccf2e07-c766-49a3-9929-1c9805b08d2a', 'name': 'volume-eccf2e07-c766-49a3-9929-1c9805b08d2a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'abee88ff-5cf5-4bf1-91e0-93b19cf30046', 'attached_at': '', 'detached_at': '', 'volume_id': 'eccf2e07-c766-49a3-9929-1c9805b08d2a', 'serial': 'eccf2e07-c766-49a3-9929-1c9805b08d2a'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 850.322715] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ad97eed-197e-47d8-9a7b-98ee560a193b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.325911] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c522b82-b4cd-4343-9e0d-c2ff385b3e76 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.352413] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e834ac0e-2cd2-4bfc-ab32-e152b0244f34 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.357519] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b551343e-80de-4aff-8900-5014f3e8cee2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.375093] env[62692]: DEBUG nova.compute.provider_tree [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.382211] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-140de09d-5d84-4a73-a337-e74a7acd54fb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.401226] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40cb2aa8-0d49-4069-9ae1-7de3d7a57ca5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.421696] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8208b7e-450d-40e9-903e-37b80dc6f3be tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] The volume has not been displaced from its original location: [datastore2] volume-eccf2e07-c766-49a3-9929-1c9805b08d2a/volume-eccf2e07-c766-49a3-9929-1c9805b08d2a.vmdk. No consolidation needed. {{(pid=62692) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 850.426959] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8208b7e-450d-40e9-903e-37b80dc6f3be tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Reconfiguring VM instance instance-00000023 to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 850.427276] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e53ed703-ecb4-478f-927c-789202afe014 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.450305] env[62692]: DEBUG oslo_vmware.api [None req-f8208b7e-450d-40e9-903e-37b80dc6f3be tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 850.450305] env[62692]: value = "task-1141176" [ 850.450305] env[62692]: _type = "Task" [ 850.450305] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.459854] env[62692]: DEBUG oslo_vmware.api [None req-f8208b7e-450d-40e9-903e-37b80dc6f3be tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1141176, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.477685] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Task: {'id': task-1141173, 'name': ReconfigVM_Task, 'duration_secs': 0.303992} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.478140] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Reconfigured VM instance instance-00000036 to attach disk [datastore2] c6a965d3-1c52-4f95-a226-9d15b7197ce7/c6a965d3-1c52-4f95-a226-9d15b7197ce7.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 850.478859] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5370d68e-2fb2-44d5-b4d0-46aaac47510d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.485456] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Waiting for the task: (returnval){ [ 850.485456] env[62692]: value = "task-1141177" [ 850.485456] env[62692]: _type = "Task" [ 850.485456] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.498831] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Task: {'id': task-1141177, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.563122] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Releasing lock "refresh_cache-18eb004f-a16b-477a-8d05-ed23a418b069" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.563122] env[62692]: DEBUG nova.compute.manager [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Instance network_info: |[{"id": "ddc1a77b-94d9-468b-ab57-8aef5343e580", "address": "fa:16:3e:4a:ca:b3", "network": {"id": "8628e5f2-19ef-4119-9620-67957a0fe952", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-559068606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa9699bac6e44e6c8132c2eb63aca648", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddc1a77b-94", "ovs_interfaceid": "ddc1a77b-94d9-468b-ab57-8aef5343e580", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 850.563122] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4a:ca:b3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ddc1a77b-94d9-468b-ab57-8aef5343e580', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 850.563122] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Creating folder: Project (fa9699bac6e44e6c8132c2eb63aca648). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 850.563122] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1a220378-682a-47e8-8c7b-9a255af0d1ed {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.563122] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Created folder: Project (fa9699bac6e44e6c8132c2eb63aca648) in parent group-v248868. [ 850.563122] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Creating folder: Instances. Parent ref: group-v248950. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 850.563122] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-20c623e2-9808-49f1-a993-0e1c9705a239 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.574979] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Created folder: Instances in parent group-v248950. [ 850.575781] env[62692]: DEBUG oslo.service.loopingcall [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 850.577777] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 850.578262] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f3d2f84a-78e8-4957-8697-390c101c88fb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.611325] env[62692]: DEBUG oslo_vmware.api [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Task: {'id': task-1141175, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.100113} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.613030] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 850.614025] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 850.614025] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 850.614025] env[62692]: INFO nova.compute.manager [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Took 1.62 seconds to destroy the instance on the hypervisor. [ 850.614025] env[62692]: DEBUG oslo.service.loopingcall [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 850.614719] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 850.614719] env[62692]: value = "task-1141180" [ 850.614719] env[62692]: _type = "Task" [ 850.614719] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.614719] env[62692]: DEBUG nova.compute.manager [-] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 850.614719] env[62692]: DEBUG nova.network.neutron [-] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 850.626300] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141180, 'name': CreateVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.636377] env[62692]: DEBUG nova.network.neutron [-] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 850.657386] env[62692]: DEBUG oslo_vmware.api [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]522f7b2d-4c01-5c6d-7a7b-e260560a24af, 'name': SearchDatastore_Task, 'duration_secs': 0.011332} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.658037] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb02810d-7d3b-4072-8ce8-c05cf3d47d37 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.663139] env[62692]: DEBUG nova.compute.manager [req-9679e9fd-3ab5-4b16-9faf-94d74743e9e2 req-2e87976c-6e4f-4c6c-8d57-fb4edcfcf450 service nova] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Received event network-changed-ddc1a77b-94d9-468b-ab57-8aef5343e580 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 850.663346] env[62692]: DEBUG nova.compute.manager [req-9679e9fd-3ab5-4b16-9faf-94d74743e9e2 req-2e87976c-6e4f-4c6c-8d57-fb4edcfcf450 service nova] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Refreshing instance network info cache due to event network-changed-ddc1a77b-94d9-468b-ab57-8aef5343e580. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 850.663565] env[62692]: DEBUG oslo_concurrency.lockutils [req-9679e9fd-3ab5-4b16-9faf-94d74743e9e2 req-2e87976c-6e4f-4c6c-8d57-fb4edcfcf450 service nova] Acquiring lock "refresh_cache-18eb004f-a16b-477a-8d05-ed23a418b069" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.663756] env[62692]: DEBUG oslo_concurrency.lockutils [req-9679e9fd-3ab5-4b16-9faf-94d74743e9e2 req-2e87976c-6e4f-4c6c-8d57-fb4edcfcf450 service nova] Acquired lock "refresh_cache-18eb004f-a16b-477a-8d05-ed23a418b069" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.663987] env[62692]: DEBUG nova.network.neutron [req-9679e9fd-3ab5-4b16-9faf-94d74743e9e2 req-2e87976c-6e4f-4c6c-8d57-fb4edcfcf450 service nova] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Refreshing network info cache for port ddc1a77b-94d9-468b-ab57-8aef5343e580 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 850.670332] env[62692]: DEBUG oslo_vmware.api [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 850.670332] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]521106a3-0a3e-e50e-c747-9c3e67a36d41" [ 850.670332] env[62692]: _type = "Task" [ 850.670332] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.686040] env[62692]: DEBUG oslo_vmware.api [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]521106a3-0a3e-e50e-c747-9c3e67a36d41, 'name': SearchDatastore_Task, 'duration_secs': 0.012572} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.686326] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.686577] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 64a8329d-e0e9-4c2d-bd1f-844ee40eb980/64a8329d-e0e9-4c2d-bd1f-844ee40eb980.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 850.686846] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ddee87fb-af98-4e7d-add8-deddd2270283 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.694131] env[62692]: DEBUG oslo_vmware.api [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 850.694131] env[62692]: value = "task-1141181" [ 850.694131] env[62692]: _type = "Task" [ 850.694131] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.703490] env[62692]: DEBUG oslo_vmware.api [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141181, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.751791] env[62692]: DEBUG nova.compute.manager [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 850.780394] env[62692]: DEBUG nova.virt.hardware [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 850.780633] env[62692]: DEBUG nova.virt.hardware [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 850.780791] env[62692]: DEBUG nova.virt.hardware [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 850.781020] env[62692]: DEBUG nova.virt.hardware [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 850.781185] env[62692]: DEBUG nova.virt.hardware [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 850.781332] env[62692]: DEBUG nova.virt.hardware [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 850.781540] env[62692]: DEBUG nova.virt.hardware [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 850.781700] env[62692]: DEBUG nova.virt.hardware [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 850.781868] env[62692]: DEBUG nova.virt.hardware [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 850.782058] env[62692]: DEBUG nova.virt.hardware [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 850.782239] env[62692]: DEBUG nova.virt.hardware [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 850.783097] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df86e3d2-e685-4271-8bc0-fba3739500db {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.790740] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ee13fda-b903-41ff-a452-02d7af4c39a2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.884034] env[62692]: DEBUG nova.scheduler.client.report [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 850.962912] env[62692]: DEBUG oslo_vmware.api [None req-f8208b7e-450d-40e9-903e-37b80dc6f3be tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1141176, 'name': ReconfigVM_Task, 'duration_secs': 0.277252} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.963215] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8208b7e-450d-40e9-903e-37b80dc6f3be tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Reconfigured VM instance instance-00000023 to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 850.971125] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1612fb25-1504-4093-9a2a-02585cd3028c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.983875] env[62692]: DEBUG oslo_vmware.api [None req-f8208b7e-450d-40e9-903e-37b80dc6f3be tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 850.983875] env[62692]: value = "task-1141182" [ 850.983875] env[62692]: _type = "Task" [ 850.983875] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.994594] env[62692]: DEBUG oslo_vmware.api [None req-f8208b7e-450d-40e9-903e-37b80dc6f3be tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1141182, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.997552] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Task: {'id': task-1141177, 'name': Rename_Task, 'duration_secs': 0.167564} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.997813] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 850.998075] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-61728a56-9103-4917-830f-a9de780b9450 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.004260] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Waiting for the task: (returnval){ [ 851.004260] env[62692]: value = "task-1141183" [ 851.004260] env[62692]: _type = "Task" [ 851.004260] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.014418] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Task: {'id': task-1141183, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.130088] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141180, 'name': CreateVM_Task, 'duration_secs': 0.35738} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.130088] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 851.130246] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.130525] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.130970] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 851.131483] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-198d8d13-d5cd-4308-9469-2c6288c460fb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.137380] env[62692]: DEBUG oslo_vmware.api [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Waiting for the task: (returnval){ [ 851.137380] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52ee0594-cc80-fa13-a597-7f3e1207bf9c" [ 851.137380] env[62692]: _type = "Task" [ 851.137380] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.141765] env[62692]: DEBUG nova.network.neutron [-] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.148871] env[62692]: DEBUG oslo_vmware.api [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52ee0594-cc80-fa13-a597-7f3e1207bf9c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.187026] env[62692]: DEBUG nova.compute.manager [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Stashing vm_state: active {{(pid=62692) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 851.207546] env[62692]: DEBUG oslo_vmware.api [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141181, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.389188] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.673s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.390606] env[62692]: DEBUG nova.compute.manager [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 851.395596] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.525s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.395596] env[62692]: DEBUG nova.objects.instance [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Lazy-loading 'resources' on Instance uuid aa0666ce-4d4f-4333-8acb-ea0cfb438cd9 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 851.410393] env[62692]: DEBUG nova.network.neutron [req-9679e9fd-3ab5-4b16-9faf-94d74743e9e2 req-2e87976c-6e4f-4c6c-8d57-fb4edcfcf450 service nova] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Updated VIF entry in instance network info cache for port ddc1a77b-94d9-468b-ab57-8aef5343e580. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 851.410393] env[62692]: DEBUG nova.network.neutron [req-9679e9fd-3ab5-4b16-9faf-94d74743e9e2 req-2e87976c-6e4f-4c6c-8d57-fb4edcfcf450 service nova] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Updating instance_info_cache with network_info: [{"id": "ddc1a77b-94d9-468b-ab57-8aef5343e580", "address": "fa:16:3e:4a:ca:b3", "network": {"id": "8628e5f2-19ef-4119-9620-67957a0fe952", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-559068606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa9699bac6e44e6c8132c2eb63aca648", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddc1a77b-94", "ovs_interfaceid": "ddc1a77b-94d9-468b-ab57-8aef5343e580", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.493770] env[62692]: DEBUG oslo_vmware.api [None req-f8208b7e-450d-40e9-903e-37b80dc6f3be tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1141182, 'name': ReconfigVM_Task, 'duration_secs': 0.165072} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.494103] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8208b7e-450d-40e9-903e-37b80dc6f3be tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-248943', 'volume_id': 'eccf2e07-c766-49a3-9929-1c9805b08d2a', 'name': 'volume-eccf2e07-c766-49a3-9929-1c9805b08d2a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'abee88ff-5cf5-4bf1-91e0-93b19cf30046', 'attached_at': '', 'detached_at': '', 'volume_id': 'eccf2e07-c766-49a3-9929-1c9805b08d2a', 'serial': 'eccf2e07-c766-49a3-9929-1c9805b08d2a'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 851.514109] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Task: {'id': task-1141183, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.648273] env[62692]: DEBUG oslo_vmware.api [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52ee0594-cc80-fa13-a597-7f3e1207bf9c, 'name': SearchDatastore_Task, 'duration_secs': 0.063668} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.648604] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.648901] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 851.649182] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.649310] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.649485] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 851.650418] env[62692]: INFO nova.compute.manager [-] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Took 1.04 seconds to deallocate network for instance. [ 851.650418] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-77fef8d8-b86d-4ca9-92be-622be46c9a7d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.659244] env[62692]: DEBUG nova.network.neutron [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Successfully updated port: a291ddc4-8996-4755-b347-066f38e6ec7d {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 851.672244] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 851.672244] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 851.672826] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f688d1ca-7066-45c6-9718-83ab20b860ba {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.683600] env[62692]: DEBUG oslo_vmware.api [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Waiting for the task: (returnval){ [ 851.683600] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52543d1c-e0a6-ca28-fba2-603d6db7c550" [ 851.683600] env[62692]: _type = "Task" [ 851.683600] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.693476] env[62692]: DEBUG oslo_vmware.api [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52543d1c-e0a6-ca28-fba2-603d6db7c550, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.704841] env[62692]: DEBUG oslo_vmware.api [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141181, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.715302} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.705733] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 64a8329d-e0e9-4c2d-bd1f-844ee40eb980/64a8329d-e0e9-4c2d-bd1f-844ee40eb980.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 851.705976] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 851.706291] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-91d6752d-ea56-461f-b372-a75025029c63 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.713597] env[62692]: DEBUG oslo_concurrency.lockutils [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.714632] env[62692]: DEBUG oslo_vmware.api [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 851.714632] env[62692]: value = "task-1141184" [ 851.714632] env[62692]: _type = "Task" [ 851.714632] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.722917] env[62692]: DEBUG oslo_vmware.api [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141184, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.897491] env[62692]: DEBUG nova.compute.utils [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 851.898986] env[62692]: DEBUG nova.compute.manager [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 851.899186] env[62692]: DEBUG nova.network.neutron [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 851.911735] env[62692]: DEBUG oslo_concurrency.lockutils [req-9679e9fd-3ab5-4b16-9faf-94d74743e9e2 req-2e87976c-6e4f-4c6c-8d57-fb4edcfcf450 service nova] Releasing lock "refresh_cache-18eb004f-a16b-477a-8d05-ed23a418b069" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.972961] env[62692]: DEBUG nova.policy [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1399b3d529cb47458ebbf234bafe9a1f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9667b0b7abaa4bf0a7e54e9d540df728', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 852.021023] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Task: {'id': task-1141183, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.078522] env[62692]: DEBUG nova.objects.instance [None req-f8208b7e-450d-40e9-903e-37b80dc6f3be tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lazy-loading 'flavor' on Instance uuid abee88ff-5cf5-4bf1-91e0-93b19cf30046 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 852.163425] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "refresh_cache-e2383a6a-3581-40fc-a0eb-6981acdbf54a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.163565] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquired lock "refresh_cache-e2383a6a-3581-40fc-a0eb-6981acdbf54a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.163713] env[62692]: DEBUG nova.network.neutron [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 852.165857] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.196117] env[62692]: DEBUG oslo_vmware.api [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52543d1c-e0a6-ca28-fba2-603d6db7c550, 'name': SearchDatastore_Task, 'duration_secs': 0.027155} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.196117] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92b66439-2a48-446e-ac00-5c39de83f3ff {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.201600] env[62692]: DEBUG oslo_vmware.api [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Waiting for the task: (returnval){ [ 852.201600] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52ca4dc4-9438-1e99-61fa-c5a8982523cc" [ 852.201600] env[62692]: _type = "Task" [ 852.201600] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.212391] env[62692]: DEBUG oslo_vmware.api [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52ca4dc4-9438-1e99-61fa-c5a8982523cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.225397] env[62692]: DEBUG oslo_vmware.api [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141184, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068408} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.225664] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 852.226472] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36cecfb2-e597-4aaa-b45b-89b9522fbd03 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.253210] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] 64a8329d-e0e9-4c2d-bd1f-844ee40eb980/64a8329d-e0e9-4c2d-bd1f-844ee40eb980.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 852.256224] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d240221-f941-4e01-8bdb-661df3267ebe {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.281600] env[62692]: DEBUG oslo_vmware.api [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 852.281600] env[62692]: value = "task-1141185" [ 852.281600] env[62692]: _type = "Task" [ 852.281600] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.291283] env[62692]: DEBUG oslo_vmware.api [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141185, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.368843] env[62692]: DEBUG nova.network.neutron [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Successfully created port: 318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 852.398749] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37eb90d4-1d4b-49ba-bb73-9e5c4504cc06 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.405731] env[62692]: DEBUG nova.compute.manager [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 852.409086] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-013380a5-4e5e-42d9-80ea-3cd3f1edc912 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.441759] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bbf561d-a047-413e-894b-d2e6e39d2ae7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.450021] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf8dbe49-ee3d-4f89-8352-335923f7a3f1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.468029] env[62692]: DEBUG nova.compute.provider_tree [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 852.517357] env[62692]: DEBUG oslo_vmware.api [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Task: {'id': task-1141183, 'name': PowerOnVM_Task, 'duration_secs': 1.318054} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.517603] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 852.517807] env[62692]: INFO nova.compute.manager [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Took 9.45 seconds to spawn the instance on the hypervisor. [ 852.517995] env[62692]: DEBUG nova.compute.manager [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 852.518835] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-041f5fde-ead9-4846-bb1a-42e19f5ca9f4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.712849] env[62692]: DEBUG oslo_vmware.api [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52ca4dc4-9438-1e99-61fa-c5a8982523cc, 'name': SearchDatastore_Task, 'duration_secs': 0.056069} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.713180] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.713398] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 18eb004f-a16b-477a-8d05-ed23a418b069/18eb004f-a16b-477a-8d05-ed23a418b069.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 852.713659] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7e0ada9f-e462-4ec6-b4c8-95bbc04431cc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.720453] env[62692]: DEBUG oslo_vmware.api [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Waiting for the task: (returnval){ [ 852.720453] env[62692]: value = "task-1141186" [ 852.720453] env[62692]: _type = "Task" [ 852.720453] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.728098] env[62692]: DEBUG oslo_vmware.api [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Task: {'id': task-1141186, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.730138] env[62692]: DEBUG nova.network.neutron [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 852.797567] env[62692]: DEBUG oslo_vmware.api [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141185, 'name': ReconfigVM_Task, 'duration_secs': 0.270852} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.797870] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Reconfigured VM instance instance-00000037 to attach disk [datastore2] 64a8329d-e0e9-4c2d-bd1f-844ee40eb980/64a8329d-e0e9-4c2d-bd1f-844ee40eb980.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 852.798520] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-274a5f75-1ae1-46af-b66c-c77d0e963fcd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.804158] env[62692]: DEBUG oslo_vmware.api [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 852.804158] env[62692]: value = "task-1141187" [ 852.804158] env[62692]: _type = "Task" [ 852.804158] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.811750] env[62692]: DEBUG oslo_vmware.api [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141187, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.824555] env[62692]: DEBUG nova.compute.manager [req-1dead275-2855-46bb-8207-9f2a8467b369 req-fbad6dd7-62af-4970-a19c-76a285461b3e service nova] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Received event network-vif-plugged-a291ddc4-8996-4755-b347-066f38e6ec7d {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 852.824843] env[62692]: DEBUG oslo_concurrency.lockutils [req-1dead275-2855-46bb-8207-9f2a8467b369 req-fbad6dd7-62af-4970-a19c-76a285461b3e service nova] Acquiring lock "e2383a6a-3581-40fc-a0eb-6981acdbf54a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.825135] env[62692]: DEBUG oslo_concurrency.lockutils [req-1dead275-2855-46bb-8207-9f2a8467b369 req-fbad6dd7-62af-4970-a19c-76a285461b3e service nova] Lock "e2383a6a-3581-40fc-a0eb-6981acdbf54a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.825358] env[62692]: DEBUG oslo_concurrency.lockutils [req-1dead275-2855-46bb-8207-9f2a8467b369 req-fbad6dd7-62af-4970-a19c-76a285461b3e service nova] Lock "e2383a6a-3581-40fc-a0eb-6981acdbf54a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.825567] env[62692]: DEBUG nova.compute.manager [req-1dead275-2855-46bb-8207-9f2a8467b369 req-fbad6dd7-62af-4970-a19c-76a285461b3e service nova] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] No waiting events found dispatching network-vif-plugged-a291ddc4-8996-4755-b347-066f38e6ec7d {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 852.825567] env[62692]: WARNING nova.compute.manager [req-1dead275-2855-46bb-8207-9f2a8467b369 req-fbad6dd7-62af-4970-a19c-76a285461b3e service nova] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Received unexpected event network-vif-plugged-a291ddc4-8996-4755-b347-066f38e6ec7d for instance with vm_state building and task_state spawning. [ 852.825910] env[62692]: DEBUG nova.compute.manager [req-1dead275-2855-46bb-8207-9f2a8467b369 req-fbad6dd7-62af-4970-a19c-76a285461b3e service nova] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Received event network-changed-a291ddc4-8996-4755-b347-066f38e6ec7d {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 852.826242] env[62692]: DEBUG nova.compute.manager [req-1dead275-2855-46bb-8207-9f2a8467b369 req-fbad6dd7-62af-4970-a19c-76a285461b3e service nova] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Refreshing instance network info cache due to event network-changed-a291ddc4-8996-4755-b347-066f38e6ec7d. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 852.826483] env[62692]: DEBUG oslo_concurrency.lockutils [req-1dead275-2855-46bb-8207-9f2a8467b369 req-fbad6dd7-62af-4970-a19c-76a285461b3e service nova] Acquiring lock "refresh_cache-e2383a6a-3581-40fc-a0eb-6981acdbf54a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.954285] env[62692]: DEBUG nova.network.neutron [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Updating instance_info_cache with network_info: [{"id": "a291ddc4-8996-4755-b347-066f38e6ec7d", "address": "fa:16:3e:7b:10:2a", "network": {"id": "2febb660-329d-403c-8c2c-f11c0f5f547c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1806650357-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37921f67664544478cb2568def068b11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa291ddc4-89", "ovs_interfaceid": "a291ddc4-8996-4755-b347-066f38e6ec7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.972410] env[62692]: DEBUG nova.scheduler.client.report [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 853.037181] env[62692]: INFO nova.compute.manager [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Took 35.49 seconds to build instance. [ 853.087555] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f8208b7e-450d-40e9-903e-37b80dc6f3be tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lock "abee88ff-5cf5-4bf1-91e0-93b19cf30046" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.352s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.231182] env[62692]: DEBUG oslo_vmware.api [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Task: {'id': task-1141186, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.321108] env[62692]: DEBUG oslo_vmware.api [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141187, 'name': Rename_Task, 'duration_secs': 0.156459} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.321453] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 853.321710] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cd15c2cf-c662-4747-b400-bc15417a9833 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.330170] env[62692]: DEBUG oslo_vmware.api [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 853.330170] env[62692]: value = "task-1141188" [ 853.330170] env[62692]: _type = "Task" [ 853.330170] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.341352] env[62692]: DEBUG oslo_vmware.api [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141188, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.423878] env[62692]: DEBUG nova.compute.manager [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 853.450798] env[62692]: DEBUG nova.virt.hardware [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 853.451118] env[62692]: DEBUG nova.virt.hardware [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 853.451340] env[62692]: DEBUG nova.virt.hardware [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 853.451554] env[62692]: DEBUG nova.virt.hardware [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 853.451740] env[62692]: DEBUG nova.virt.hardware [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 853.451912] env[62692]: DEBUG nova.virt.hardware [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 853.452166] env[62692]: DEBUG nova.virt.hardware [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 853.452338] env[62692]: DEBUG nova.virt.hardware [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 853.452525] env[62692]: DEBUG nova.virt.hardware [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 853.452706] env[62692]: DEBUG nova.virt.hardware [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 853.452897] env[62692]: DEBUG nova.virt.hardware [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 853.453850] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91133c1f-ce95-4ee3-8141-c63684f5ab79 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.457472] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Releasing lock "refresh_cache-e2383a6a-3581-40fc-a0eb-6981acdbf54a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.457472] env[62692]: DEBUG nova.compute.manager [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Instance network_info: |[{"id": "a291ddc4-8996-4755-b347-066f38e6ec7d", "address": "fa:16:3e:7b:10:2a", "network": {"id": "2febb660-329d-403c-8c2c-f11c0f5f547c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1806650357-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37921f67664544478cb2568def068b11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa291ddc4-89", "ovs_interfaceid": "a291ddc4-8996-4755-b347-066f38e6ec7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 853.457746] env[62692]: DEBUG oslo_concurrency.lockutils [req-1dead275-2855-46bb-8207-9f2a8467b369 req-fbad6dd7-62af-4970-a19c-76a285461b3e service nova] Acquired lock "refresh_cache-e2383a6a-3581-40fc-a0eb-6981acdbf54a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.458920] env[62692]: DEBUG nova.network.neutron [req-1dead275-2855-46bb-8207-9f2a8467b369 req-fbad6dd7-62af-4970-a19c-76a285461b3e service nova] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Refreshing network info cache for port a291ddc4-8996-4755-b347-066f38e6ec7d {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 853.462660] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7b:10:2a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a291ddc4-8996-4755-b347-066f38e6ec7d', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 853.467200] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Creating folder: Project (37921f67664544478cb2568def068b11). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 853.471033] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ffba69cd-df62-452a-aca4-401e07d8376d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.476553] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb8e203b-b1b9-431f-a051-6a4274618e24 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.481256] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.087s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.485764] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 22.974s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.485764] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.485764] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62692) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 853.485764] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.974s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.487135] env[62692]: INFO nova.compute.claims [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 853.490724] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Created folder: Project (37921f67664544478cb2568def068b11) in parent group-v248868. [ 853.491671] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Creating folder: Instances. Parent ref: group-v248953. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 853.492114] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c07b9f28-437a-4a7d-91dd-4ba7718e266e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.495127] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-17a90e52-b728-4f9e-b8af-1087ed191ab2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.509884] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26dec2bb-38ae-49a3-a418-253c75a64fbc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.515805] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Created folder: Instances in parent group-v248953. [ 853.515805] env[62692]: DEBUG oslo.service.loopingcall [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 853.516388] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 853.517351] env[62692]: INFO nova.scheduler.client.report [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Deleted allocations for instance aa0666ce-4d4f-4333-8acb-ea0cfb438cd9 [ 853.518375] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ee8ca6e1-25dc-4aaf-a25f-d57d754692ae {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.548307] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c410ee22-6eba-43da-9258-a8504e2d9dce tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Lock "c6a965d3-1c52-4f95-a226-9d15b7197ce7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.314s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.551076] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4c0cb27-6fe8-4adf-8f4e-7b2df9108320 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.555372] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 853.555372] env[62692]: value = "task-1141191" [ 853.555372] env[62692]: _type = "Task" [ 853.555372] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.562612] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a35f11a9-b95f-4d42-bb05-a43b41c5a667 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.569391] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141191, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.596626] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180303MB free_disk=166GB free_vcpus=48 pci_devices=None {{(pid=62692) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 853.596827] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.732307] env[62692]: DEBUG oslo_vmware.api [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Task: {'id': task-1141186, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.76138} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.732581] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 18eb004f-a16b-477a-8d05-ed23a418b069/18eb004f-a16b-477a-8d05-ed23a418b069.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 853.732798] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 853.733104] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-08624923-4558-44b6-bf70-b5ea33954b5f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.740164] env[62692]: DEBUG oslo_vmware.api [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Waiting for the task: (returnval){ [ 853.740164] env[62692]: value = "task-1141192" [ 853.740164] env[62692]: _type = "Task" [ 853.740164] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.753971] env[62692]: DEBUG oslo_vmware.api [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Task: {'id': task-1141192, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.845308] env[62692]: DEBUG oslo_vmware.api [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141188, 'name': PowerOnVM_Task, 'duration_secs': 0.507787} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.845658] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 853.845896] env[62692]: INFO nova.compute.manager [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Took 8.21 seconds to spawn the instance on the hypervisor. [ 853.846182] env[62692]: DEBUG nova.compute.manager [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 853.846903] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5323e681-9509-454d-b523-5af90c952b87 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.856699] env[62692]: DEBUG nova.network.neutron [req-1dead275-2855-46bb-8207-9f2a8467b369 req-fbad6dd7-62af-4970-a19c-76a285461b3e service nova] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Updated VIF entry in instance network info cache for port a291ddc4-8996-4755-b347-066f38e6ec7d. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 853.857090] env[62692]: DEBUG nova.network.neutron [req-1dead275-2855-46bb-8207-9f2a8467b369 req-fbad6dd7-62af-4970-a19c-76a285461b3e service nova] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Updating instance_info_cache with network_info: [{"id": "a291ddc4-8996-4755-b347-066f38e6ec7d", "address": "fa:16:3e:7b:10:2a", "network": {"id": "2febb660-329d-403c-8c2c-f11c0f5f547c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1806650357-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37921f67664544478cb2568def068b11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa291ddc4-89", "ovs_interfaceid": "a291ddc4-8996-4755-b347-066f38e6ec7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.902311] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Acquiring lock "ae44f074-a5c8-4259-99c6-9ce290b1570b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.902545] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lock "ae44f074-a5c8-4259-99c6-9ce290b1570b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.020706] env[62692]: DEBUG nova.network.neutron [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Successfully updated port: 318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 854.052727] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96e786f1-7738-48e6-a51a-f3f95ada12a9 tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Lock "aa0666ce-4d4f-4333-8acb-ea0cfb438cd9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.874s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.057099] env[62692]: DEBUG nova.compute.manager [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 854.069113] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141191, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.250707] env[62692]: DEBUG oslo_vmware.api [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Task: {'id': task-1141192, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071761} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.250707] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 854.251797] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63d0b55f-e126-47dd-a443-4a5c1ab77fb0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.274281] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] 18eb004f-a16b-477a-8d05-ed23a418b069/18eb004f-a16b-477a-8d05-ed23a418b069.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 854.274637] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a52b38f6-be6e-4f03-9ca3-278439c79351 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.295434] env[62692]: DEBUG oslo_vmware.api [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Waiting for the task: (returnval){ [ 854.295434] env[62692]: value = "task-1141193" [ 854.295434] env[62692]: _type = "Task" [ 854.295434] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.303483] env[62692]: DEBUG oslo_vmware.api [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Task: {'id': task-1141193, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.363735] env[62692]: DEBUG oslo_concurrency.lockutils [req-1dead275-2855-46bb-8207-9f2a8467b369 req-fbad6dd7-62af-4970-a19c-76a285461b3e service nova] Releasing lock "refresh_cache-e2383a6a-3581-40fc-a0eb-6981acdbf54a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.366953] env[62692]: INFO nova.compute.manager [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Took 35.10 seconds to build instance. [ 854.522827] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Acquiring lock "refresh_cache-6b2d3385-a9af-432b-a1e7-c91adb2e1f69" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.523146] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Acquired lock "refresh_cache-6b2d3385-a9af-432b-a1e7-c91adb2e1f69" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.523363] env[62692]: DEBUG nova.network.neutron [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 854.572383] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141191, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.593335] env[62692]: DEBUG oslo_concurrency.lockutils [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.808963] env[62692]: DEBUG oslo_vmware.api [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Task: {'id': task-1141193, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.869517] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7b5e298a-33c7-4d1d-809c-47990a4ea7be tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "64a8329d-e0e9-4c2d-bd1f-844ee40eb980" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.581s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.969127] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bbbb445-04fd-450b-9681-3841c6970316 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.976739] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-592f47aa-1cb1-4464-b619-3f84d0b64ef9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.015942] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3858670-3c9c-46d1-a589-d7e2569857d1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.027550] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acf91453-826f-41b7-9858-8351fcf3aaab {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.048880] env[62692]: DEBUG nova.compute.provider_tree [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 855.068179] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141191, 'name': CreateVM_Task, 'duration_secs': 1.389694} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.068179] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 855.069076] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.069766] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.070294] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 855.070557] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-746b0a21-7a2a-4858-804b-e1e9503fc81a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.075755] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 855.075755] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]528e6ab5-b41a-79bf-ce5b-f6d1247b1dcc" [ 855.075755] env[62692]: _type = "Task" [ 855.075755] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.084495] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]528e6ab5-b41a-79bf-ce5b-f6d1247b1dcc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.087714] env[62692]: DEBUG nova.network.neutron [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 855.306607] env[62692]: DEBUG oslo_vmware.api [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Task: {'id': task-1141193, 'name': ReconfigVM_Task, 'duration_secs': 0.610538} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.307320] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Reconfigured VM instance instance-00000038 to attach disk [datastore2] 18eb004f-a16b-477a-8d05-ed23a418b069/18eb004f-a16b-477a-8d05-ed23a418b069.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 855.308169] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e5adb1fd-8fde-4303-b9f3-405feff2c7b6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.318491] env[62692]: DEBUG oslo_vmware.api [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Waiting for the task: (returnval){ [ 855.318491] env[62692]: value = "task-1141194" [ 855.318491] env[62692]: _type = "Task" [ 855.318491] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.330038] env[62692]: DEBUG oslo_vmware.api [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Task: {'id': task-1141194, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.375595] env[62692]: DEBUG nova.compute.manager [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 855.392958] env[62692]: DEBUG nova.network.neutron [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Updating instance_info_cache with network_info: [{"id": "318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db", "address": "fa:16:3e:32:9a:8d", "network": {"id": "7e7a1dcd-36ba-4a25-affa-6bd7fe895ade", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-508011613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9667b0b7abaa4bf0a7e54e9d540df728", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap318dc9c9-0a", "ovs_interfaceid": "318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.460495] env[62692]: DEBUG nova.compute.manager [req-5638fe11-614f-4b60-9f96-e1d31b765b9a req-139ec45c-4372-4756-9b4d-0f3721b9c6c4 service nova] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Received event network-vif-plugged-318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 855.460729] env[62692]: DEBUG oslo_concurrency.lockutils [req-5638fe11-614f-4b60-9f96-e1d31b765b9a req-139ec45c-4372-4756-9b4d-0f3721b9c6c4 service nova] Acquiring lock "6b2d3385-a9af-432b-a1e7-c91adb2e1f69-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.462744] env[62692]: DEBUG oslo_concurrency.lockutils [req-5638fe11-614f-4b60-9f96-e1d31b765b9a req-139ec45c-4372-4756-9b4d-0f3721b9c6c4 service nova] Lock "6b2d3385-a9af-432b-a1e7-c91adb2e1f69-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.462744] env[62692]: DEBUG oslo_concurrency.lockutils [req-5638fe11-614f-4b60-9f96-e1d31b765b9a req-139ec45c-4372-4756-9b4d-0f3721b9c6c4 service nova] Lock "6b2d3385-a9af-432b-a1e7-c91adb2e1f69-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.462744] env[62692]: DEBUG nova.compute.manager [req-5638fe11-614f-4b60-9f96-e1d31b765b9a req-139ec45c-4372-4756-9b4d-0f3721b9c6c4 service nova] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] No waiting events found dispatching network-vif-plugged-318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 855.463121] env[62692]: WARNING nova.compute.manager [req-5638fe11-614f-4b60-9f96-e1d31b765b9a req-139ec45c-4372-4756-9b4d-0f3721b9c6c4 service nova] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Received unexpected event network-vif-plugged-318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db for instance with vm_state building and task_state spawning. [ 855.463121] env[62692]: DEBUG nova.compute.manager [req-5638fe11-614f-4b60-9f96-e1d31b765b9a req-139ec45c-4372-4756-9b4d-0f3721b9c6c4 service nova] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Received event network-changed-318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 855.463250] env[62692]: DEBUG nova.compute.manager [req-5638fe11-614f-4b60-9f96-e1d31b765b9a req-139ec45c-4372-4756-9b4d-0f3721b9c6c4 service nova] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Refreshing instance network info cache due to event network-changed-318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 855.463417] env[62692]: DEBUG oslo_concurrency.lockutils [req-5638fe11-614f-4b60-9f96-e1d31b765b9a req-139ec45c-4372-4756-9b4d-0f3721b9c6c4 service nova] Acquiring lock "refresh_cache-6b2d3385-a9af-432b-a1e7-c91adb2e1f69" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.553089] env[62692]: DEBUG nova.scheduler.client.report [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 855.590536] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]528e6ab5-b41a-79bf-ce5b-f6d1247b1dcc, 'name': SearchDatastore_Task, 'duration_secs': 0.012955} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.590816] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.591068] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 855.591306] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.591461] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.591639] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 855.592063] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-425e1b5c-13ae-4007-8604-16838994a79a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.600808] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 855.601472] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 855.601905] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f497846-4a20-4737-a061-a6a1c58357e9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.609511] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 855.609511] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52edfbff-8313-7a5f-a5d4-b214983de409" [ 855.609511] env[62692]: _type = "Task" [ 855.609511] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.618205] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52edfbff-8313-7a5f-a5d4-b214983de409, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.827371] env[62692]: DEBUG oslo_vmware.api [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Task: {'id': task-1141194, 'name': Rename_Task, 'duration_secs': 0.333545} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.827799] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 855.828172] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bf13a82c-1390-441a-843d-babfcd1cd15b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.835043] env[62692]: DEBUG oslo_vmware.api [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Waiting for the task: (returnval){ [ 855.835043] env[62692]: value = "task-1141195" [ 855.835043] env[62692]: _type = "Task" [ 855.835043] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.844133] env[62692]: DEBUG oslo_vmware.api [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Task: {'id': task-1141195, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.900016] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Releasing lock "refresh_cache-6b2d3385-a9af-432b-a1e7-c91adb2e1f69" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.900016] env[62692]: DEBUG nova.compute.manager [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Instance network_info: |[{"id": "318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db", "address": "fa:16:3e:32:9a:8d", "network": {"id": "7e7a1dcd-36ba-4a25-affa-6bd7fe895ade", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-508011613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9667b0b7abaa4bf0a7e54e9d540df728", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap318dc9c9-0a", "ovs_interfaceid": "318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 855.900016] env[62692]: DEBUG oslo_concurrency.lockutils [req-5638fe11-614f-4b60-9f96-e1d31b765b9a req-139ec45c-4372-4756-9b4d-0f3721b9c6c4 service nova] Acquired lock "refresh_cache-6b2d3385-a9af-432b-a1e7-c91adb2e1f69" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.900016] env[62692]: DEBUG nova.network.neutron [req-5638fe11-614f-4b60-9f96-e1d31b765b9a req-139ec45c-4372-4756-9b4d-0f3721b9c6c4 service nova] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Refreshing network info cache for port 318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 855.900016] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:32:9a:8d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4a8f5363-be3a-4f92-9ccf-33bb0c8113b3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 855.912623] env[62692]: DEBUG oslo.service.loopingcall [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 855.914022] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.915013] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 855.915394] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-39476f0b-7fdf-48e0-8699-d64de9dadf02 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.935160] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 855.935160] env[62692]: value = "task-1141196" [ 855.935160] env[62692]: _type = "Task" [ 855.935160] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.943077] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141196, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.031020] env[62692]: DEBUG nova.compute.manager [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 856.031965] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1760c99a-dde4-42ee-90cd-8cc4827064c9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.062514] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.577s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.063264] env[62692]: DEBUG nova.compute.manager [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 856.066641] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 23.048s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.066906] env[62692]: DEBUG nova.objects.instance [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62692) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 856.121018] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52edfbff-8313-7a5f-a5d4-b214983de409, 'name': SearchDatastore_Task, 'duration_secs': 0.009592} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.121670] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca5a40e8-9457-4c67-b5a2-a2871eed40ee {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.126684] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 856.126684] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52e716d9-e648-547f-a523-86e31c55ab28" [ 856.126684] env[62692]: _type = "Task" [ 856.126684] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.134568] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52e716d9-e648-547f-a523-86e31c55ab28, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.349920] env[62692]: DEBUG oslo_vmware.api [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Task: {'id': task-1141195, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.446883] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141196, 'name': CreateVM_Task, 'duration_secs': 0.366213} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.447086] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 856.447726] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.447896] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.448235] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 856.448493] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10b72533-1757-41e4-b164-8b82331b8b2c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.455764] env[62692]: DEBUG oslo_vmware.api [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Waiting for the task: (returnval){ [ 856.455764] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]529e3668-700a-ce06-3f02-13ae55c1b164" [ 856.455764] env[62692]: _type = "Task" [ 856.455764] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.463709] env[62692]: DEBUG oslo_vmware.api [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]529e3668-700a-ce06-3f02-13ae55c1b164, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.543150] env[62692]: INFO nova.compute.manager [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] instance snapshotting [ 856.550833] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5148604c-a025-4c18-89f7-6b304003871c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.573718] env[62692]: DEBUG nova.compute.utils [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 856.580568] env[62692]: DEBUG nova.compute.manager [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 856.580881] env[62692]: DEBUG nova.network.neutron [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 856.584867] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cac8bbf-93b5-42d6-9e8c-54814c209fde {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.592093] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5ec5a8c0-3a0f-4a14-973b-669f4418504e tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 0.525s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.593468] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.816s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.593718] env[62692]: DEBUG nova.objects.instance [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Lazy-loading 'resources' on Instance uuid 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 856.638135] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52e716d9-e648-547f-a523-86e31c55ab28, 'name': SearchDatastore_Task, 'duration_secs': 0.009138} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.638416] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.638678] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] e2383a6a-3581-40fc-a0eb-6981acdbf54a/e2383a6a-3581-40fc-a0eb-6981acdbf54a.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 856.639020] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2c0c47df-9403-4ea0-935d-23252537ae5d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.647694] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 856.647694] env[62692]: value = "task-1141197" [ 856.647694] env[62692]: _type = "Task" [ 856.647694] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.655646] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141197, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.670234] env[62692]: DEBUG nova.policy [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b0c1314a9aed42e5b54a33de96f3db1f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '71b00c566cea4b21811ea1e1ef5ba0f3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 856.722385] env[62692]: DEBUG nova.network.neutron [req-5638fe11-614f-4b60-9f96-e1d31b765b9a req-139ec45c-4372-4756-9b4d-0f3721b9c6c4 service nova] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Updated VIF entry in instance network info cache for port 318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 856.722385] env[62692]: DEBUG nova.network.neutron [req-5638fe11-614f-4b60-9f96-e1d31b765b9a req-139ec45c-4372-4756-9b4d-0f3721b9c6c4 service nova] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Updating instance_info_cache with network_info: [{"id": "318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db", "address": "fa:16:3e:32:9a:8d", "network": {"id": "7e7a1dcd-36ba-4a25-affa-6bd7fe895ade", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-508011613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9667b0b7abaa4bf0a7e54e9d540df728", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap318dc9c9-0a", "ovs_interfaceid": "318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.845341] env[62692]: DEBUG oslo_vmware.api [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Task: {'id': task-1141195, 'name': PowerOnVM_Task, 'duration_secs': 0.890212} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.845600] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 856.845811] env[62692]: INFO nova.compute.manager [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Took 8.66 seconds to spawn the instance on the hypervisor. [ 856.845994] env[62692]: DEBUG nova.compute.manager [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 856.846782] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c58dfb9f-e2f4-4212-b97c-220d6a544523 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.969886] env[62692]: DEBUG oslo_vmware.api [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]529e3668-700a-ce06-3f02-13ae55c1b164, 'name': SearchDatastore_Task, 'duration_secs': 0.011599} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.969886] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.969886] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 856.970241] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.970439] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.970625] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 856.970880] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-842a445a-0a2f-4dc9-8235-c957b3f55943 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.980095] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 856.980319] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 856.981052] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24939d90-648c-4f28-8b6a-d730c2bc39bc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.986318] env[62692]: DEBUG oslo_vmware.api [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Waiting for the task: (returnval){ [ 856.986318] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52b26132-5c77-3c6e-64f4-dc9cb508023c" [ 856.986318] env[62692]: _type = "Task" [ 856.986318] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.994313] env[62692]: DEBUG oslo_vmware.api [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52b26132-5c77-3c6e-64f4-dc9cb508023c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.079999] env[62692]: DEBUG nova.compute.manager [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 857.099166] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "interface-27583cb4-0478-44d8-8970-478682271513-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.099455] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "interface-27583cb4-0478-44d8-8970-478682271513-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.099825] env[62692]: DEBUG nova.objects.instance [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lazy-loading 'flavor' on Instance uuid 27583cb4-0478-44d8-8970-478682271513 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 857.109206] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Creating Snapshot of the VM instance {{(pid=62692) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 857.109206] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-7354c182-2473-4609-9374-9f833ded7bd8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.118234] env[62692]: DEBUG oslo_vmware.api [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 857.118234] env[62692]: value = "task-1141198" [ 857.118234] env[62692]: _type = "Task" [ 857.118234] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.131585] env[62692]: DEBUG oslo_vmware.api [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141198, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.153550] env[62692]: DEBUG nova.network.neutron [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Successfully created port: 152b902d-a99b-44d9-974f-71348e613fc9 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 857.163283] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141197, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.224810] env[62692]: DEBUG oslo_concurrency.lockutils [req-5638fe11-614f-4b60-9f96-e1d31b765b9a req-139ec45c-4372-4756-9b4d-0f3721b9c6c4 service nova] Releasing lock "refresh_cache-6b2d3385-a9af-432b-a1e7-c91adb2e1f69" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.366376] env[62692]: INFO nova.compute.manager [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Took 33.16 seconds to build instance. [ 857.498390] env[62692]: DEBUG oslo_vmware.api [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52b26132-5c77-3c6e-64f4-dc9cb508023c, 'name': SearchDatastore_Task, 'duration_secs': 0.02188} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.502399] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aab56ba2-81c6-4cf9-872e-61c0d0a9f83a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.508300] env[62692]: DEBUG oslo_vmware.api [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Waiting for the task: (returnval){ [ 857.508300] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]528cd112-f321-ad29-1408-7da06ce18ce7" [ 857.508300] env[62692]: _type = "Task" [ 857.508300] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.516844] env[62692]: DEBUG oslo_vmware.api [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]528cd112-f321-ad29-1408-7da06ce18ce7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.596332] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37ce97b7-be36-4877-8cc2-7835aaa8e21c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.604264] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87102808-098c-48bf-a890-7de961d60f34 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.641954] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6e42930-c671-418f-a74d-7ff7453b7eb2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.652654] env[62692]: DEBUG oslo_vmware.api [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141198, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.659283] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-278b88a0-c66b-4201-b379-f7525a0e158b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.670975] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141197, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.784097} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.678948] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] e2383a6a-3581-40fc-a0eb-6981acdbf54a/e2383a6a-3581-40fc-a0eb-6981acdbf54a.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 857.679837] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 857.680201] env[62692]: DEBUG nova.compute.provider_tree [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.681086] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5dfcbd4a-0c72-4c88-ab7d-96c8a627a1fa {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.687524] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 857.687524] env[62692]: value = "task-1141199" [ 857.687524] env[62692]: _type = "Task" [ 857.687524] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.696697] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141199, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.710889] env[62692]: DEBUG nova.objects.instance [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lazy-loading 'pci_requests' on Instance uuid 27583cb4-0478-44d8-8970-478682271513 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 857.869617] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d7136f7e-135c-4f86-a06c-2084e5b6cc5c tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Lock "18eb004f-a16b-477a-8d05-ed23a418b069" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 96.827s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.018740] env[62692]: DEBUG oslo_vmware.api [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]528cd112-f321-ad29-1408-7da06ce18ce7, 'name': SearchDatastore_Task, 'duration_secs': 0.016501} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.019531] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.019531] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 6b2d3385-a9af-432b-a1e7-c91adb2e1f69/6b2d3385-a9af-432b-a1e7-c91adb2e1f69.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 858.020367] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-014e681a-2c27-4f05-8c24-6c98c95ff6d9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.026000] env[62692]: DEBUG oslo_vmware.api [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Waiting for the task: (returnval){ [ 858.026000] env[62692]: value = "task-1141200" [ 858.026000] env[62692]: _type = "Task" [ 858.026000] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.034254] env[62692]: DEBUG oslo_vmware.api [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1141200, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.095186] env[62692]: DEBUG nova.compute.manager [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 858.126225] env[62692]: DEBUG nova.virt.hardware [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 858.126631] env[62692]: DEBUG nova.virt.hardware [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 858.126704] env[62692]: DEBUG nova.virt.hardware [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 858.126885] env[62692]: DEBUG nova.virt.hardware [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 858.127042] env[62692]: DEBUG nova.virt.hardware [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 858.127198] env[62692]: DEBUG nova.virt.hardware [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 858.127409] env[62692]: DEBUG nova.virt.hardware [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 858.127567] env[62692]: DEBUG nova.virt.hardware [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 858.127733] env[62692]: DEBUG nova.virt.hardware [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 858.127963] env[62692]: DEBUG nova.virt.hardware [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 858.128089] env[62692]: DEBUG nova.virt.hardware [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 858.129151] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f3ef80e-91ae-4b08-9da7-abe0fcaf6bac {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.136795] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81cd3567-c8f4-4cc8-a2bb-3d5299ed59bd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.157890] env[62692]: DEBUG oslo_vmware.api [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141198, 'name': CreateSnapshot_Task, 'duration_secs': 1.029975} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.158213] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Created Snapshot of the VM instance {{(pid=62692) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 858.158992] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-039ef34e-3247-4531-837d-57d186f17427 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.184820] env[62692]: DEBUG nova.scheduler.client.report [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 858.198497] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141199, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.222289} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.198792] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 858.199656] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3f1bd9e-9912-4248-9862-332a6ddc1e30 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.218403] env[62692]: DEBUG nova.objects.base [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Object Instance<27583cb4-0478-44d8-8970-478682271513> lazy-loaded attributes: flavor,pci_requests {{(pid=62692) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 858.218896] env[62692]: DEBUG nova.network.neutron [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 858.230711] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] e2383a6a-3581-40fc-a0eb-6981acdbf54a/e2383a6a-3581-40fc-a0eb-6981acdbf54a.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 858.232484] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd91b3ec-050f-4142-a7c8-3d314beeaa36 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.251686] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Acquiring lock "9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.251953] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Lock "9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.259385] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 858.259385] env[62692]: value = "task-1141201" [ 858.259385] env[62692]: _type = "Task" [ 858.259385] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.268502] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141201, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.296152] env[62692]: DEBUG nova.policy [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a251ef5f6cb143178928babc75a0fef0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ba97a775e99a463c845518b0df8cd3e2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 858.543305] env[62692]: DEBUG oslo_vmware.api [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1141200, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.620023] env[62692]: DEBUG nova.network.neutron [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Successfully created port: 1b690180-b6c8-4a42-96de-cfd6f25d1791 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 858.678037] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Creating linked-clone VM from snapshot {{(pid=62692) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 858.678465] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-c34e5a53-ae06-4248-a8f6-48623db553c5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.687605] env[62692]: DEBUG oslo_vmware.api [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 858.687605] env[62692]: value = "task-1141202" [ 858.687605] env[62692]: _type = "Task" [ 858.687605] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.696244] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.103s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.698294] env[62692]: DEBUG oslo_vmware.api [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141202, 'name': CloneVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.698800] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.498s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.700375] env[62692]: INFO nova.compute.claims [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 858.728456] env[62692]: INFO nova.scheduler.client.report [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Deleted allocations for instance 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99 [ 858.756771] env[62692]: DEBUG nova.compute.manager [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 858.773774] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141201, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.041074] env[62692]: DEBUG oslo_vmware.api [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1141200, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.574811} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.041074] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 6b2d3385-a9af-432b-a1e7-c91adb2e1f69/6b2d3385-a9af-432b-a1e7-c91adb2e1f69.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 859.041074] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 859.041074] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f9f9b471-f323-4b78-b9e2-2ee0e3d05274 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.047255] env[62692]: DEBUG oslo_vmware.api [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Waiting for the task: (returnval){ [ 859.047255] env[62692]: value = "task-1141203" [ 859.047255] env[62692]: _type = "Task" [ 859.047255] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.059798] env[62692]: DEBUG oslo_vmware.api [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1141203, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.155348] env[62692]: DEBUG nova.network.neutron [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Successfully updated port: 152b902d-a99b-44d9-974f-71348e613fc9 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 859.160978] env[62692]: DEBUG nova.compute.manager [req-cd2f7db8-c331-41c0-ae7e-fdaa32a3c250 req-86b7e27e-d25f-4df2-bc95-5368dfea3be7 service nova] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Received event network-changed-fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 859.162028] env[62692]: DEBUG nova.compute.manager [req-cd2f7db8-c331-41c0-ae7e-fdaa32a3c250 req-86b7e27e-d25f-4df2-bc95-5368dfea3be7 service nova] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Refreshing instance network info cache due to event network-changed-fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 859.162028] env[62692]: DEBUG oslo_concurrency.lockutils [req-cd2f7db8-c331-41c0-ae7e-fdaa32a3c250 req-86b7e27e-d25f-4df2-bc95-5368dfea3be7 service nova] Acquiring lock "refresh_cache-c6a965d3-1c52-4f95-a226-9d15b7197ce7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.162028] env[62692]: DEBUG oslo_concurrency.lockutils [req-cd2f7db8-c331-41c0-ae7e-fdaa32a3c250 req-86b7e27e-d25f-4df2-bc95-5368dfea3be7 service nova] Acquired lock "refresh_cache-c6a965d3-1c52-4f95-a226-9d15b7197ce7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.162028] env[62692]: DEBUG nova.network.neutron [req-cd2f7db8-c331-41c0-ae7e-fdaa32a3c250 req-86b7e27e-d25f-4df2-bc95-5368dfea3be7 service nova] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Refreshing network info cache for port fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 859.198560] env[62692]: DEBUG oslo_vmware.api [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141202, 'name': CloneVM_Task} progress is 94%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.239747] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e7cb403f-7f8b-4667-ab81-25bfdbdafc9f tempest-ServerGroupTestJSON-1066639624 tempest-ServerGroupTestJSON-1066639624-project-member] Lock "1cdd1dbc-567d-46f6-9f05-dfbc57acbf99" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.229s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.271945] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141201, 'name': ReconfigVM_Task, 'duration_secs': 0.515911} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.274224] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Reconfigured VM instance instance-00000039 to attach disk [datastore2] e2383a6a-3581-40fc-a0eb-6981acdbf54a/e2383a6a-3581-40fc-a0eb-6981acdbf54a.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 859.275105] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0f8756be-fbe3-4277-9de6-f266c174c814 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.284137] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 859.284137] env[62692]: value = "task-1141204" [ 859.284137] env[62692]: _type = "Task" [ 859.284137] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.285134] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.294120] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141204, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.558364] env[62692]: DEBUG oslo_vmware.api [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1141203, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087059} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.558675] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 859.559485] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8281bdb9-d27c-4401-9a66-7f5db6968bb3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.583178] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] 6b2d3385-a9af-432b-a1e7-c91adb2e1f69/6b2d3385-a9af-432b-a1e7-c91adb2e1f69.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 859.583595] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9329e89c-c1f5-4fa3-86a8-db12a0898bfe {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.607119] env[62692]: DEBUG oslo_vmware.api [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Waiting for the task: (returnval){ [ 859.607119] env[62692]: value = "task-1141205" [ 859.607119] env[62692]: _type = "Task" [ 859.607119] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.616146] env[62692]: DEBUG oslo_vmware.api [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1141205, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.658053] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "refresh_cache-754b3c65-1e4b-49d2-8980-095d975edb01" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.658189] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquired lock "refresh_cache-754b3c65-1e4b-49d2-8980-095d975edb01" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.658339] env[62692]: DEBUG nova.network.neutron [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 859.698976] env[62692]: DEBUG oslo_vmware.api [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141202, 'name': CloneVM_Task} progress is 94%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.798653] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141204, 'name': Rename_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.058565] env[62692]: DEBUG nova.network.neutron [req-cd2f7db8-c331-41c0-ae7e-fdaa32a3c250 req-86b7e27e-d25f-4df2-bc95-5368dfea3be7 service nova] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Updated VIF entry in instance network info cache for port fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 860.059083] env[62692]: DEBUG nova.network.neutron [req-cd2f7db8-c331-41c0-ae7e-fdaa32a3c250 req-86b7e27e-d25f-4df2-bc95-5368dfea3be7 service nova] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Updating instance_info_cache with network_info: [{"id": "fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8", "address": "fa:16:3e:c7:7a:a0", "network": {"id": "04aacc3d-37b6-4a5c-ac4d-55313731e4f9", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1367536055-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef45ad1482dc4f5db51a5badc3cdce22", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a0734cc4-5718-45e2-9f98-0ded96880bef", "external-id": "nsx-vlan-transportzone-875", "segmentation_id": 875, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdd5ae79-fe", "ovs_interfaceid": "fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.120289] env[62692]: DEBUG oslo_vmware.api [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1141205, 'name': ReconfigVM_Task, 'duration_secs': 0.510011} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.123172] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Reconfigured VM instance instance-0000003a to attach disk [datastore1] 6b2d3385-a9af-432b-a1e7-c91adb2e1f69/6b2d3385-a9af-432b-a1e7-c91adb2e1f69.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 860.124076] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d3228358-0938-4222-8a39-71da7314166c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.133393] env[62692]: DEBUG oslo_vmware.api [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Waiting for the task: (returnval){ [ 860.133393] env[62692]: value = "task-1141206" [ 860.133393] env[62692]: _type = "Task" [ 860.133393] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.146166] env[62692]: DEBUG oslo_vmware.api [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1141206, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.166894] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dd97cbc-fe06-4980-a353-3d10dee4f854 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.174583] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3baf03dc-d43c-4c79-b1a0-43e97e1ada2c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.208079] env[62692]: DEBUG nova.network.neutron [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 860.213804] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae0744f7-3d53-49d1-946d-79b0012d08ed {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.220283] env[62692]: DEBUG nova.network.neutron [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Successfully updated port: 1b690180-b6c8-4a42-96de-cfd6f25d1791 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 860.226329] env[62692]: DEBUG oslo_vmware.api [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141202, 'name': CloneVM_Task} progress is 94%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.227820] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67e62b67-420c-489b-ab64-0882d2b7f23d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.246634] env[62692]: DEBUG nova.compute.provider_tree [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 860.296369] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141204, 'name': Rename_Task, 'duration_secs': 0.626558} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.296642] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 860.296895] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4c21fa91-44a0-4bf8-b5af-b5761e21350c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.303256] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 860.303256] env[62692]: value = "task-1141207" [ 860.303256] env[62692]: _type = "Task" [ 860.303256] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.313957] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141207, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.430557] env[62692]: DEBUG nova.network.neutron [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Updating instance_info_cache with network_info: [{"id": "152b902d-a99b-44d9-974f-71348e613fc9", "address": "fa:16:3e:11:96:8e", "network": {"id": "e9c538a3-7723-42cb-85b7-5a1597e49cef", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1878627765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71b00c566cea4b21811ea1e1ef5ba0f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap152b902d-a9", "ovs_interfaceid": "152b902d-a99b-44d9-974f-71348e613fc9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.481976] env[62692]: DEBUG nova.compute.manager [req-33e0f229-b72e-46cc-ad0e-874725379c01 req-0be58fba-2539-4521-811b-c439fd3c3954 service nova] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Received event network-vif-plugged-152b902d-a99b-44d9-974f-71348e613fc9 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 860.481976] env[62692]: DEBUG oslo_concurrency.lockutils [req-33e0f229-b72e-46cc-ad0e-874725379c01 req-0be58fba-2539-4521-811b-c439fd3c3954 service nova] Acquiring lock "754b3c65-1e4b-49d2-8980-095d975edb01-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.481976] env[62692]: DEBUG oslo_concurrency.lockutils [req-33e0f229-b72e-46cc-ad0e-874725379c01 req-0be58fba-2539-4521-811b-c439fd3c3954 service nova] Lock "754b3c65-1e4b-49d2-8980-095d975edb01-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.481976] env[62692]: DEBUG oslo_concurrency.lockutils [req-33e0f229-b72e-46cc-ad0e-874725379c01 req-0be58fba-2539-4521-811b-c439fd3c3954 service nova] Lock "754b3c65-1e4b-49d2-8980-095d975edb01-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.482204] env[62692]: DEBUG nova.compute.manager [req-33e0f229-b72e-46cc-ad0e-874725379c01 req-0be58fba-2539-4521-811b-c439fd3c3954 service nova] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] No waiting events found dispatching network-vif-plugged-152b902d-a99b-44d9-974f-71348e613fc9 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 860.482204] env[62692]: WARNING nova.compute.manager [req-33e0f229-b72e-46cc-ad0e-874725379c01 req-0be58fba-2539-4521-811b-c439fd3c3954 service nova] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Received unexpected event network-vif-plugged-152b902d-a99b-44d9-974f-71348e613fc9 for instance with vm_state building and task_state spawning. [ 860.562606] env[62692]: DEBUG oslo_concurrency.lockutils [req-cd2f7db8-c331-41c0-ae7e-fdaa32a3c250 req-86b7e27e-d25f-4df2-bc95-5368dfea3be7 service nova] Releasing lock "refresh_cache-c6a965d3-1c52-4f95-a226-9d15b7197ce7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.645403] env[62692]: DEBUG oslo_vmware.api [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1141206, 'name': Rename_Task, 'duration_secs': 0.163836} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.645690] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 860.645963] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-59a3abde-f02f-4368-ac67-a78551328823 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.654382] env[62692]: DEBUG oslo_vmware.api [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Waiting for the task: (returnval){ [ 860.654382] env[62692]: value = "task-1141208" [ 860.654382] env[62692]: _type = "Task" [ 860.654382] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.664053] env[62692]: DEBUG oslo_vmware.api [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1141208, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.716939] env[62692]: DEBUG oslo_vmware.api [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141202, 'name': CloneVM_Task} progress is 95%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.728042] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "refresh_cache-27583cb4-0478-44d8-8970-478682271513" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.729041] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "refresh_cache-27583cb4-0478-44d8-8970-478682271513" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.729041] env[62692]: DEBUG nova.network.neutron [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 860.751654] env[62692]: DEBUG nova.scheduler.client.report [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 860.816461] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141207, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.934994] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Releasing lock "refresh_cache-754b3c65-1e4b-49d2-8980-095d975edb01" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.935371] env[62692]: DEBUG nova.compute.manager [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Instance network_info: |[{"id": "152b902d-a99b-44d9-974f-71348e613fc9", "address": "fa:16:3e:11:96:8e", "network": {"id": "e9c538a3-7723-42cb-85b7-5a1597e49cef", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1878627765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71b00c566cea4b21811ea1e1ef5ba0f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap152b902d-a9", "ovs_interfaceid": "152b902d-a99b-44d9-974f-71348e613fc9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 860.935822] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:11:96:8e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dd96b39f-bd2e-48d1-85c3-577cf97f08c8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '152b902d-a99b-44d9-974f-71348e613fc9', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 860.943715] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Creating folder: Project (71b00c566cea4b21811ea1e1ef5ba0f3). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 860.944053] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e71fa476-07f6-4d23-9f56-c51a53122000 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.957285] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Created folder: Project (71b00c566cea4b21811ea1e1ef5ba0f3) in parent group-v248868. [ 860.957496] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Creating folder: Instances. Parent ref: group-v248959. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 860.957738] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f42f1a6b-fe6f-4a14-99b1-601f5e9d6d48 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.969409] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Created folder: Instances in parent group-v248959. [ 860.969714] env[62692]: DEBUG oslo.service.loopingcall [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 860.969992] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 860.970332] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7ae1b9c7-a3e4-46f8-92de-2e982ccfe42d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.992141] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 860.992141] env[62692]: value = "task-1141211" [ 860.992141] env[62692]: _type = "Task" [ 860.992141] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.000594] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141211, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.165573] env[62692]: DEBUG oslo_vmware.api [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1141208, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.217627] env[62692]: DEBUG oslo_vmware.api [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141202, 'name': CloneVM_Task, 'duration_secs': 2.186825} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.217910] env[62692]: INFO nova.virt.vmwareapi.vmops [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Created linked-clone VM from snapshot [ 861.218704] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a97faec5-f102-4505-ae65-d48370c999ba {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.227139] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Uploading image 427c2d72-be26-42e4-9f8c-ca628806da41 {{(pid=62692) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 861.254005] env[62692]: DEBUG oslo_vmware.rw_handles [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 861.254005] env[62692]: value = "vm-248958" [ 861.254005] env[62692]: _type = "VirtualMachine" [ 861.254005] env[62692]: }. {{(pid=62692) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 861.254325] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-d9125afd-4d68-4834-8665-49a970ba7518 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.258017] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.559s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.258552] env[62692]: DEBUG nova.compute.manager [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 861.263049] env[62692]: DEBUG oslo_concurrency.lockutils [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 23.923s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.263458] env[62692]: DEBUG nova.objects.instance [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62692) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 861.266317] env[62692]: DEBUG oslo_vmware.rw_handles [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lease: (returnval){ [ 861.266317] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5247b09a-999c-23a2-d7fc-428d96e317d1" [ 861.266317] env[62692]: _type = "HttpNfcLease" [ 861.266317] env[62692]: } obtained for exporting VM: (result){ [ 861.266317] env[62692]: value = "vm-248958" [ 861.266317] env[62692]: _type = "VirtualMachine" [ 861.266317] env[62692]: }. {{(pid=62692) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 861.266910] env[62692]: DEBUG oslo_vmware.api [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the lease: (returnval){ [ 861.266910] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5247b09a-999c-23a2-d7fc-428d96e317d1" [ 861.266910] env[62692]: _type = "HttpNfcLease" [ 861.266910] env[62692]: } to be ready. {{(pid=62692) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 861.274486] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 861.274486] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5247b09a-999c-23a2-d7fc-428d96e317d1" [ 861.274486] env[62692]: _type = "HttpNfcLease" [ 861.274486] env[62692]: } is initializing. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 861.295332] env[62692]: WARNING nova.network.neutron [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] b34af32c-59b7-4de4-85ef-08bd67c1e2a2 already exists in list: networks containing: ['b34af32c-59b7-4de4-85ef-08bd67c1e2a2']. ignoring it [ 861.314733] env[62692]: DEBUG oslo_vmware.api [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141207, 'name': PowerOnVM_Task, 'duration_secs': 0.530968} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.314996] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 861.315252] env[62692]: INFO nova.compute.manager [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Took 10.56 seconds to spawn the instance on the hypervisor. [ 861.315443] env[62692]: DEBUG nova.compute.manager [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 861.316939] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c94bbf0d-d923-49fa-83dc-0757f8e54f93 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.503431] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141211, 'name': CreateVM_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.608874] env[62692]: INFO nova.compute.manager [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Rebuilding instance [ 861.654847] env[62692]: DEBUG nova.compute.manager [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 861.655741] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0ef0435-1727-4e47-b3c5-ca56ea1acd04 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.674099] env[62692]: DEBUG oslo_vmware.api [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1141208, 'name': PowerOnVM_Task, 'duration_secs': 0.802201} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.674580] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 861.674784] env[62692]: INFO nova.compute.manager [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Took 8.25 seconds to spawn the instance on the hypervisor. [ 861.674990] env[62692]: DEBUG nova.compute.manager [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 861.675715] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a20dc0e-cea1-4bcc-9e83-c56adf6e5c67 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.768876] env[62692]: DEBUG nova.compute.utils [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 861.772949] env[62692]: DEBUG nova.compute.manager [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 861.773142] env[62692]: DEBUG nova.network.neutron [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 861.783227] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 861.783227] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5247b09a-999c-23a2-d7fc-428d96e317d1" [ 861.783227] env[62692]: _type = "HttpNfcLease" [ 861.783227] env[62692]: } is ready. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 861.786621] env[62692]: DEBUG oslo_vmware.rw_handles [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 861.786621] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5247b09a-999c-23a2-d7fc-428d96e317d1" [ 861.786621] env[62692]: _type = "HttpNfcLease" [ 861.786621] env[62692]: }. {{(pid=62692) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 861.786621] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0961552a-31de-4018-8cc5-ed457aeced6c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.795671] env[62692]: DEBUG oslo_vmware.rw_handles [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d91c56-689f-1ca1-03bd-93bd6a3add50/disk-0.vmdk from lease info. {{(pid=62692) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 861.796351] env[62692]: DEBUG oslo_vmware.rw_handles [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d91c56-689f-1ca1-03bd-93bd6a3add50/disk-0.vmdk for reading. {{(pid=62692) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 861.867167] env[62692]: DEBUG nova.network.neutron [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Updating instance_info_cache with network_info: [{"id": "d4991635-5c62-4473-9661-97ac75a187fb", "address": "fa:16:3e:d4:98:43", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4991635-5c", "ovs_interfaceid": "d4991635-5c62-4473-9661-97ac75a187fb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "1b690180-b6c8-4a42-96de-cfd6f25d1791", "address": "fa:16:3e:a4:db:b9", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b690180-b6", "ovs_interfaceid": "1b690180-b6c8-4a42-96de-cfd6f25d1791", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.869404] env[62692]: DEBUG nova.policy [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9e9f256c1ebf4b1ba966edb7b1df5f4f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '266c39fe585d41079d0278ecd1d1c6a4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 861.875602] env[62692]: INFO nova.compute.manager [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Took 36.80 seconds to build instance. [ 861.935135] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c7c6869e-e1d4-418a-91c4-5118c296c91b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.004526] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141211, 'name': CreateVM_Task, 'duration_secs': 0.716087} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.005502] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 862.005502] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.005502] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.005814] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 862.006083] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ee5e40a-1680-4cf3-92f6-ed51dc136898 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.011286] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 862.011286] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]526b4407-0163-226d-6ac3-3b986dc379b7" [ 862.011286] env[62692]: _type = "Task" [ 862.011286] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.019808] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]526b4407-0163-226d-6ac3-3b986dc379b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.176278] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 862.176810] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e2cccfd5-e0c3-4401-a73f-3b7c2b54ce1d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.179868] env[62692]: DEBUG nova.network.neutron [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Successfully created port: 513b63b4-6498-4af7-b62f-9bdbc4f1e9ed {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 862.198751] env[62692]: DEBUG oslo_vmware.api [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Waiting for the task: (returnval){ [ 862.198751] env[62692]: value = "task-1141213" [ 862.198751] env[62692]: _type = "Task" [ 862.198751] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.199380] env[62692]: INFO nova.compute.manager [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Took 34.48 seconds to build instance. [ 862.211658] env[62692]: DEBUG oslo_vmware.api [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Task: {'id': task-1141213, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.274197] env[62692]: DEBUG nova.compute.manager [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 862.279940] env[62692]: DEBUG oslo_concurrency.lockutils [None req-389ca362-3f4f-4a49-8bc7-4f6eb582e7a6 tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.284869] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.814s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.284869] env[62692]: DEBUG nova.objects.instance [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Lazy-loading 'resources' on Instance uuid 1848c251-0a0c-4c36-8bd0-ff37befde2a0 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 862.375606] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "refresh_cache-27583cb4-0478-44d8-8970-478682271513" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.376326] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "27583cb4-0478-44d8-8970-478682271513" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.376606] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "27583cb4-0478-44d8-8970-478682271513" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.377690] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51a7cd3f-71f9-493d-b9c8-973ef5821aa0 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "e2383a6a-3581-40fc-a0eb-6981acdbf54a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.205s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.378719] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56e4f2fc-ee1c-407b-b7f3-766cce15ade5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.404035] env[62692]: DEBUG nova.virt.hardware [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 862.404035] env[62692]: DEBUG nova.virt.hardware [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 862.404144] env[62692]: DEBUG nova.virt.hardware [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 862.404518] env[62692]: DEBUG nova.virt.hardware [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 862.404749] env[62692]: DEBUG nova.virt.hardware [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 862.405016] env[62692]: DEBUG nova.virt.hardware [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 862.405264] env[62692]: DEBUG nova.virt.hardware [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 862.405437] env[62692]: DEBUG nova.virt.hardware [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 862.405625] env[62692]: DEBUG nova.virt.hardware [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 862.407550] env[62692]: DEBUG nova.virt.hardware [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 862.407550] env[62692]: DEBUG nova.virt.hardware [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 862.414047] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Reconfiguring VM to attach interface {{(pid=62692) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 862.414522] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd9d651b-bb7e-4fd2-a6f1-8771acfeccc1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.435076] env[62692]: DEBUG oslo_vmware.api [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 862.435076] env[62692]: value = "task-1141214" [ 862.435076] env[62692]: _type = "Task" [ 862.435076] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.444635] env[62692]: DEBUG oslo_vmware.api [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141214, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.524461] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]526b4407-0163-226d-6ac3-3b986dc379b7, 'name': SearchDatastore_Task, 'duration_secs': 0.010563} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.524820] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.525082] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 862.525711] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.525711] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.525711] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 862.525995] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0163a907-1e78-4f8a-bb04-fa1b30f05592 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.539983] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 862.539983] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 862.539983] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-547c10fd-6cc1-404a-ba26-a49d1eca2073 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.545776] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 862.545776] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52ce3f81-dbd9-3720-820c-3ebc85d882f9" [ 862.545776] env[62692]: _type = "Task" [ 862.545776] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.555444] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52ce3f81-dbd9-3720-820c-3ebc85d882f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.706014] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28ee75a2-4289-4233-aa34-fcafe9158843 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Lock "6b2d3385-a9af-432b-a1e7-c91adb2e1f69" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.305s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.712361] env[62692]: DEBUG oslo_vmware.api [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Task: {'id': task-1141213, 'name': PowerOffVM_Task, 'duration_secs': 0.234973} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.712706] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 862.713667] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 862.713977] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e4c7f1d7-232f-4790-9da1-b3d2c30ca3fb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.723982] env[62692]: DEBUG oslo_vmware.api [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Waiting for the task: (returnval){ [ 862.723982] env[62692]: value = "task-1141215" [ 862.723982] env[62692]: _type = "Task" [ 862.723982] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.733727] env[62692]: DEBUG oslo_vmware.api [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Task: {'id': task-1141215, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.949471] env[62692]: DEBUG oslo_vmware.api [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141214, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.057706] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52ce3f81-dbd9-3720-820c-3ebc85d882f9, 'name': SearchDatastore_Task, 'duration_secs': 0.016041} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.061529] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca6e1fe0-503e-41c1-be28-05153ac7a886 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.067972] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 863.067972] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52a5eb7c-c1f3-32a4-d0d5-68dfbaa1e1ae" [ 863.067972] env[62692]: _type = "Task" [ 863.067972] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.076218] env[62692]: DEBUG nova.compute.manager [req-9cd58490-7b0c-49d9-bc2a-b2fa2f3c937c req-d1699098-0c9c-42ae-aba6-b82a6730e1bf service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] Received event network-vif-plugged-1b690180-b6c8-4a42-96de-cfd6f25d1791 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 863.076572] env[62692]: DEBUG oslo_concurrency.lockutils [req-9cd58490-7b0c-49d9-bc2a-b2fa2f3c937c req-d1699098-0c9c-42ae-aba6-b82a6730e1bf service nova] Acquiring lock "27583cb4-0478-44d8-8970-478682271513-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.076892] env[62692]: DEBUG oslo_concurrency.lockutils [req-9cd58490-7b0c-49d9-bc2a-b2fa2f3c937c req-d1699098-0c9c-42ae-aba6-b82a6730e1bf service nova] Lock "27583cb4-0478-44d8-8970-478682271513-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.077110] env[62692]: DEBUG oslo_concurrency.lockutils [req-9cd58490-7b0c-49d9-bc2a-b2fa2f3c937c req-d1699098-0c9c-42ae-aba6-b82a6730e1bf service nova] Lock "27583cb4-0478-44d8-8970-478682271513-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.077340] env[62692]: DEBUG nova.compute.manager [req-9cd58490-7b0c-49d9-bc2a-b2fa2f3c937c req-d1699098-0c9c-42ae-aba6-b82a6730e1bf service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] No waiting events found dispatching network-vif-plugged-1b690180-b6c8-4a42-96de-cfd6f25d1791 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 863.077521] env[62692]: WARNING nova.compute.manager [req-9cd58490-7b0c-49d9-bc2a-b2fa2f3c937c req-d1699098-0c9c-42ae-aba6-b82a6730e1bf service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] Received unexpected event network-vif-plugged-1b690180-b6c8-4a42-96de-cfd6f25d1791 for instance with vm_state active and task_state None. [ 863.077748] env[62692]: DEBUG nova.compute.manager [req-9cd58490-7b0c-49d9-bc2a-b2fa2f3c937c req-d1699098-0c9c-42ae-aba6-b82a6730e1bf service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] Received event network-changed-1b690180-b6c8-4a42-96de-cfd6f25d1791 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 863.077924] env[62692]: DEBUG nova.compute.manager [req-9cd58490-7b0c-49d9-bc2a-b2fa2f3c937c req-d1699098-0c9c-42ae-aba6-b82a6730e1bf service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] Refreshing instance network info cache due to event network-changed-1b690180-b6c8-4a42-96de-cfd6f25d1791. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 863.078241] env[62692]: DEBUG oslo_concurrency.lockutils [req-9cd58490-7b0c-49d9-bc2a-b2fa2f3c937c req-d1699098-0c9c-42ae-aba6-b82a6730e1bf service nova] Acquiring lock "refresh_cache-27583cb4-0478-44d8-8970-478682271513" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.078487] env[62692]: DEBUG oslo_concurrency.lockutils [req-9cd58490-7b0c-49d9-bc2a-b2fa2f3c937c req-d1699098-0c9c-42ae-aba6-b82a6730e1bf service nova] Acquired lock "refresh_cache-27583cb4-0478-44d8-8970-478682271513" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.079223] env[62692]: DEBUG nova.network.neutron [req-9cd58490-7b0c-49d9-bc2a-b2fa2f3c937c req-d1699098-0c9c-42ae-aba6-b82a6730e1bf service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] Refreshing network info cache for port 1b690180-b6c8-4a42-96de-cfd6f25d1791 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 863.087813] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52a5eb7c-c1f3-32a4-d0d5-68dfbaa1e1ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.199181] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd41526-c5df-4cad-b03e-cfdb6509dfd0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.208014] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f940395-2de3-405c-9170-8bd5c343003a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.246386] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90f5ac9b-e4d4-4262-a2eb-6c7a42115db3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.258548] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] VM already powered off {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 863.258874] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Volume detach. Driver type: vmdk {{(pid=62692) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 863.259139] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-248875', 'volume_id': '9886d0c7-5df2-4bce-8853-faa1431b1865', 'name': 'volume-9886d0c7-5df2-4bce-8853-faa1431b1865', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '81ce340c-fbef-4932-983c-595843530dbc', 'attached_at': '', 'detached_at': '', 'volume_id': '9886d0c7-5df2-4bce-8853-faa1431b1865', 'serial': '9886d0c7-5df2-4bce-8853-faa1431b1865'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 863.260046] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00884d66-6db1-4d3f-9a1d-56e94f5ffdf9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.263991] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b754368b-0982-44ed-aaa6-479d63489cba {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.279948] env[62692]: DEBUG nova.compute.provider_tree [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 863.298018] env[62692]: DEBUG nova.compute.manager [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 863.300034] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39eac163-7c15-477d-b52b-10e8b704f263 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.309338] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04c1c579-4f31-4d88-bc7b-cccee25e9ac5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.337097] env[62692]: DEBUG nova.virt.hardware [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 863.337354] env[62692]: DEBUG nova.virt.hardware [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 863.337570] env[62692]: DEBUG nova.virt.hardware [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 863.337867] env[62692]: DEBUG nova.virt.hardware [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 863.338117] env[62692]: DEBUG nova.virt.hardware [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 863.338294] env[62692]: DEBUG nova.virt.hardware [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 863.338589] env[62692]: DEBUG nova.virt.hardware [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 863.338775] env[62692]: DEBUG nova.virt.hardware [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 863.339036] env[62692]: DEBUG nova.virt.hardware [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 863.339310] env[62692]: DEBUG nova.virt.hardware [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 863.339524] env[62692]: DEBUG nova.virt.hardware [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 863.340553] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3504202c-629a-44b2-8053-7cbef1e04596 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.343943] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4572892a-d3ec-4a10-a35f-9397396b4a9e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.365200] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c05dff5-cdc9-4b3b-ac61-11b8acb64e23 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.369566] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] The volume has not been displaced from its original location: [datastore2] volume-9886d0c7-5df2-4bce-8853-faa1431b1865/volume-9886d0c7-5df2-4bce-8853-faa1431b1865.vmdk. No consolidation needed. {{(pid=62692) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 863.375076] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Reconfiguring VM instance instance-0000002f to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 863.375790] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-88f6fb77-dbe8-4214-aa50-8d3aeef5fc1b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.403734] env[62692]: DEBUG oslo_vmware.api [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Waiting for the task: (returnval){ [ 863.403734] env[62692]: value = "task-1141216" [ 863.403734] env[62692]: _type = "Task" [ 863.403734] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.413915] env[62692]: DEBUG oslo_vmware.api [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Task: {'id': task-1141216, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.446690] env[62692]: DEBUG oslo_vmware.api [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141214, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.585391] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52a5eb7c-c1f3-32a4-d0d5-68dfbaa1e1ae, 'name': SearchDatastore_Task, 'duration_secs': 0.024987} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.585693] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.586064] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 754b3c65-1e4b-49d2-8980-095d975edb01/754b3c65-1e4b-49d2-8980-095d975edb01.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 863.586378] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d235b8f8-cb69-463c-af63-10863a4df605 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.597754] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 863.597754] env[62692]: value = "task-1141217" [ 863.597754] env[62692]: _type = "Task" [ 863.597754] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.610935] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141217, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.788657] env[62692]: DEBUG nova.network.neutron [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Successfully updated port: 513b63b4-6498-4af7-b62f-9bdbc4f1e9ed {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 863.799637] env[62692]: DEBUG nova.scheduler.client.report [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 863.915860] env[62692]: DEBUG oslo_vmware.api [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Task: {'id': task-1141216, 'name': ReconfigVM_Task, 'duration_secs': 0.234063} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.916209] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Reconfigured VM instance instance-0000002f to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 863.922034] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0df845e7-8395-4cd9-88cc-3912e4b85a79 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.936583] env[62692]: DEBUG nova.network.neutron [req-9cd58490-7b0c-49d9-bc2a-b2fa2f3c937c req-d1699098-0c9c-42ae-aba6-b82a6730e1bf service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] Updated VIF entry in instance network info cache for port 1b690180-b6c8-4a42-96de-cfd6f25d1791. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 863.937088] env[62692]: DEBUG nova.network.neutron [req-9cd58490-7b0c-49d9-bc2a-b2fa2f3c937c req-d1699098-0c9c-42ae-aba6-b82a6730e1bf service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] Updating instance_info_cache with network_info: [{"id": "d4991635-5c62-4473-9661-97ac75a187fb", "address": "fa:16:3e:d4:98:43", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4991635-5c", "ovs_interfaceid": "d4991635-5c62-4473-9661-97ac75a187fb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "1b690180-b6c8-4a42-96de-cfd6f25d1791", "address": "fa:16:3e:a4:db:b9", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b690180-b6", "ovs_interfaceid": "1b690180-b6c8-4a42-96de-cfd6f25d1791", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.943894] env[62692]: DEBUG oslo_vmware.api [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Waiting for the task: (returnval){ [ 863.943894] env[62692]: value = "task-1141218" [ 863.943894] env[62692]: _type = "Task" [ 863.943894] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.951584] env[62692]: DEBUG oslo_vmware.api [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141214, 'name': ReconfigVM_Task, 'duration_secs': 1.344404} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.953330] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "27583cb4-0478-44d8-8970-478682271513" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.953575] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Reconfigured VM to attach interface {{(pid=62692) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 863.960363] env[62692]: DEBUG oslo_vmware.api [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Task: {'id': task-1141218, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.111856] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141217, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.295241] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "refresh_cache-26c1ff5b-9eda-4131-a4ea-d4511a2364b7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.295241] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquired lock "refresh_cache-26c1ff5b-9eda-4131-a4ea-d4511a2364b7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.295241] env[62692]: DEBUG nova.network.neutron [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 864.309291] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.028s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.313316] env[62692]: DEBUG nova.compute.manager [req-92c780b4-dfa9-4e9e-a60d-b03e1a53c40f req-60e4313d-74ae-48a6-86fd-2b51a16135a4 service nova] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Received event network-changed-152b902d-a99b-44d9-974f-71348e613fc9 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 864.313541] env[62692]: DEBUG nova.compute.manager [req-92c780b4-dfa9-4e9e-a60d-b03e1a53c40f req-60e4313d-74ae-48a6-86fd-2b51a16135a4 service nova] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Refreshing instance network info cache due to event network-changed-152b902d-a99b-44d9-974f-71348e613fc9. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 864.313762] env[62692]: DEBUG oslo_concurrency.lockutils [req-92c780b4-dfa9-4e9e-a60d-b03e1a53c40f req-60e4313d-74ae-48a6-86fd-2b51a16135a4 service nova] Acquiring lock "refresh_cache-754b3c65-1e4b-49d2-8980-095d975edb01" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.313953] env[62692]: DEBUG oslo_concurrency.lockutils [req-92c780b4-dfa9-4e9e-a60d-b03e1a53c40f req-60e4313d-74ae-48a6-86fd-2b51a16135a4 service nova] Acquired lock "refresh_cache-754b3c65-1e4b-49d2-8980-095d975edb01" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.314175] env[62692]: DEBUG nova.network.neutron [req-92c780b4-dfa9-4e9e-a60d-b03e1a53c40f req-60e4313d-74ae-48a6-86fd-2b51a16135a4 service nova] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Refreshing network info cache for port 152b902d-a99b-44d9-974f-71348e613fc9 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 864.317102] env[62692]: DEBUG oslo_concurrency.lockutils [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.143s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.317102] env[62692]: DEBUG nova.objects.instance [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lazy-loading 'resources' on Instance uuid b1d6409a-2733-470f-a929-672fe1631b1b {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 864.381585] env[62692]: INFO nova.scheduler.client.report [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Deleted allocations for instance 1848c251-0a0c-4c36-8bd0-ff37befde2a0 [ 864.445353] env[62692]: DEBUG oslo_concurrency.lockutils [req-9cd58490-7b0c-49d9-bc2a-b2fa2f3c937c req-d1699098-0c9c-42ae-aba6-b82a6730e1bf service nova] Releasing lock "refresh_cache-27583cb4-0478-44d8-8970-478682271513" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.457388] env[62692]: DEBUG oslo_vmware.api [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Task: {'id': task-1141218, 'name': ReconfigVM_Task, 'duration_secs': 0.169622} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.457729] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-248875', 'volume_id': '9886d0c7-5df2-4bce-8853-faa1431b1865', 'name': 'volume-9886d0c7-5df2-4bce-8853-faa1431b1865', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '81ce340c-fbef-4932-983c-595843530dbc', 'attached_at': '', 'detached_at': '', 'volume_id': '9886d0c7-5df2-4bce-8853-faa1431b1865', 'serial': '9886d0c7-5df2-4bce-8853-faa1431b1865'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 864.458032] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 864.458897] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d4138d-8b39-4422-956d-cd157e0ee3d0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.462759] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b6d1ba4b-1af1-4b33-98b9-d1529752fe74 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "interface-27583cb4-0478-44d8-8970-478682271513-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.363s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.471600] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 864.471895] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b2e274e5-4360-4185-a0ee-6d7c668b4450 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.569443] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 864.569688] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 864.569906] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Deleting the datastore file [datastore2] 81ce340c-fbef-4932-983c-595843530dbc {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 864.570286] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f2b92608-7513-4047-9db9-200f947f8211 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.581889] env[62692]: DEBUG oslo_vmware.api [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Waiting for the task: (returnval){ [ 864.581889] env[62692]: value = "task-1141220" [ 864.581889] env[62692]: _type = "Task" [ 864.581889] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.593530] env[62692]: DEBUG oslo_vmware.api [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Task: {'id': task-1141220, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.609034] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141217, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.703862} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.609400] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 754b3c65-1e4b-49d2-8980-095d975edb01/754b3c65-1e4b-49d2-8980-095d975edb01.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 864.609632] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 864.609946] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-100541d2-9393-4b38-8cba-bff536051ff2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.620231] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 864.620231] env[62692]: value = "task-1141221" [ 864.620231] env[62692]: _type = "Task" [ 864.620231] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.630156] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141221, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.844276] env[62692]: DEBUG nova.network.neutron [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 864.892914] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7ff1d8d9-d7d9-4341-b3a6-2cd6619b6f09 tempest-ServerShowV257Test-1315000434 tempest-ServerShowV257Test-1315000434-project-member] Lock "1848c251-0a0c-4c36-8bd0-ff37befde2a0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.171s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.101371] env[62692]: DEBUG oslo_vmware.api [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Task: {'id': task-1141220, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137996} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.101371] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 865.101556] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 865.101717] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 865.111392] env[62692]: DEBUG nova.network.neutron [req-92c780b4-dfa9-4e9e-a60d-b03e1a53c40f req-60e4313d-74ae-48a6-86fd-2b51a16135a4 service nova] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Updated VIF entry in instance network info cache for port 152b902d-a99b-44d9-974f-71348e613fc9. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 865.111746] env[62692]: DEBUG nova.network.neutron [req-92c780b4-dfa9-4e9e-a60d-b03e1a53c40f req-60e4313d-74ae-48a6-86fd-2b51a16135a4 service nova] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Updating instance_info_cache with network_info: [{"id": "152b902d-a99b-44d9-974f-71348e613fc9", "address": "fa:16:3e:11:96:8e", "network": {"id": "e9c538a3-7723-42cb-85b7-5a1597e49cef", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1878627765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71b00c566cea4b21811ea1e1ef5ba0f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap152b902d-a9", "ovs_interfaceid": "152b902d-a99b-44d9-974f-71348e613fc9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.140277] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141221, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091079} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.145779] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 865.147386] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e91fa1e5-34cd-43f6-be79-18e066475094 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.175576] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] 754b3c65-1e4b-49d2-8980-095d975edb01/754b3c65-1e4b-49d2-8980-095d975edb01.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 865.178915] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c6a77a93-4f10-4719-bcbe-0d8a981aa84b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.194578] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Volume detach. Driver type: vmdk {{(pid=62692) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 865.196345] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-38ea0b2b-c7a4-401d-b86d-f059de0678de {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.212341] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d11c7097-15b7-48dc-bd34-1b18e1fe854b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.223710] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 865.223710] env[62692]: value = "task-1141222" [ 865.223710] env[62692]: _type = "Task" [ 865.223710] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.256796] env[62692]: ERROR nova.compute.manager [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Failed to detach volume 9886d0c7-5df2-4bce-8853-faa1431b1865 from /dev/sda: nova.exception.InstanceNotFound: Instance 81ce340c-fbef-4932-983c-595843530dbc could not be found. [ 865.256796] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] Traceback (most recent call last): [ 865.256796] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 865.256796] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] self.driver.rebuild(**kwargs) [ 865.256796] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] File "/opt/stack/nova/nova/virt/driver.py", line 390, in rebuild [ 865.256796] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] raise NotImplementedError() [ 865.256796] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] NotImplementedError [ 865.256796] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] [ 865.256796] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] During handling of the above exception, another exception occurred: [ 865.256796] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] [ 865.256796] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] Traceback (most recent call last): [ 865.256796] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 865.256796] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] self.driver.detach_volume(context, old_connection_info, [ 865.256796] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 556, in detach_volume [ 865.256796] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] return self._volumeops.detach_volume(connection_info, instance) [ 865.256796] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 865.256796] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] self._detach_volume_vmdk(connection_info, instance) [ 865.256796] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 865.256796] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 865.256796] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 865.256796] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] stable_ref.fetch_moref(session) [ 865.256796] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 865.256796] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] raise exception.InstanceNotFound(instance_id=self._uuid) [ 865.256796] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] nova.exception.InstanceNotFound: Instance 81ce340c-fbef-4932-983c-595843530dbc could not be found. [ 865.256796] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] [ 865.303567] env[62692]: DEBUG nova.network.neutron [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Updating instance_info_cache with network_info: [{"id": "513b63b4-6498-4af7-b62f-9bdbc4f1e9ed", "address": "fa:16:3e:49:d0:ff", "network": {"id": "0e06e04a-5213-4d06-b3f1-b2b8f437e6d0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1533469056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "266c39fe585d41079d0278ecd1d1c6a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "678ebbe4-4c53-4eaf-a689-93981310f37d", "external-id": "nsx-vlan-transportzone-443", "segmentation_id": 443, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap513b63b4-64", "ovs_interfaceid": "513b63b4-6498-4af7-b62f-9bdbc4f1e9ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.359298] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e48581d3-2a17-4ac3-af5f-909e3bd5c43f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.368977] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b83edc5-711a-4e3b-b4ac-2a04b199ee08 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.409141] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-397001d9-6263-4650-9287-545720e11ea6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.417898] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-229f68a3-6079-46ee-bc14-8beb81892bb7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.434191] env[62692]: DEBUG nova.compute.provider_tree [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 865.478042] env[62692]: DEBUG nova.compute.utils [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Build of instance 81ce340c-fbef-4932-983c-595843530dbc aborted: Failed to rebuild volume backed instance. {{(pid=62692) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 865.480816] env[62692]: ERROR nova.compute.manager [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance 81ce340c-fbef-4932-983c-595843530dbc aborted: Failed to rebuild volume backed instance. [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] Traceback (most recent call last): [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] self.driver.rebuild(**kwargs) [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] File "/opt/stack/nova/nova/virt/driver.py", line 390, in rebuild [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] raise NotImplementedError() [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] NotImplementedError [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] During handling of the above exception, another exception occurred: [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] Traceback (most recent call last): [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] File "/opt/stack/nova/nova/compute/manager.py", line 3600, in _rebuild_volume_backed_instance [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] self._detach_root_volume(context, instance, root_bdm) [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] File "/opt/stack/nova/nova/compute/manager.py", line 3579, in _detach_root_volume [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] with excutils.save_and_reraise_exception(): [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] self.force_reraise() [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] raise self.value [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] self.driver.detach_volume(context, old_connection_info, [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 556, in detach_volume [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] return self._volumeops.detach_volume(connection_info, instance) [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] self._detach_volume_vmdk(connection_info, instance) [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] stable_ref.fetch_moref(session) [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] raise exception.InstanceNotFound(instance_id=self._uuid) [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] nova.exception.InstanceNotFound: Instance 81ce340c-fbef-4932-983c-595843530dbc could not be found. [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] During handling of the above exception, another exception occurred: [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] Traceback (most recent call last): [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] File "/opt/stack/nova/nova/compute/manager.py", line 10863, in _error_out_instance_on_exception [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] yield [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] File "/opt/stack/nova/nova/compute/manager.py", line 3868, in rebuild_instance [ 865.480816] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] self._do_rebuild_instance_with_claim( [ 865.482400] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] File "/opt/stack/nova/nova/compute/manager.py", line 3954, in _do_rebuild_instance_with_claim [ 865.482400] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] self._do_rebuild_instance( [ 865.482400] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] File "/opt/stack/nova/nova/compute/manager.py", line 4146, in _do_rebuild_instance [ 865.482400] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] self._rebuild_default_impl(**kwargs) [ 865.482400] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] File "/opt/stack/nova/nova/compute/manager.py", line 3723, in _rebuild_default_impl [ 865.482400] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] self._rebuild_volume_backed_instance( [ 865.482400] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] File "/opt/stack/nova/nova/compute/manager.py", line 3615, in _rebuild_volume_backed_instance [ 865.482400] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] raise exception.BuildAbortException( [ 865.482400] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] nova.exception.BuildAbortException: Build of instance 81ce340c-fbef-4932-983c-595843530dbc aborted: Failed to rebuild volume backed instance. [ 865.482400] env[62692]: ERROR nova.compute.manager [instance: 81ce340c-fbef-4932-983c-595843530dbc] [ 865.524563] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a4950060-a599-4fa1-a856-e9bc0c94e03a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "d99737d8-2eb0-40ee-b61e-6c736c84ea59" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.524709] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a4950060-a599-4fa1-a856-e9bc0c94e03a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "d99737d8-2eb0-40ee-b61e-6c736c84ea59" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.524826] env[62692]: INFO nova.compute.manager [None req-a4950060-a599-4fa1-a856-e9bc0c94e03a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Rebooting instance [ 865.619147] env[62692]: DEBUG oslo_concurrency.lockutils [req-92c780b4-dfa9-4e9e-a60d-b03e1a53c40f req-60e4313d-74ae-48a6-86fd-2b51a16135a4 service nova] Releasing lock "refresh_cache-754b3c65-1e4b-49d2-8980-095d975edb01" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.619147] env[62692]: DEBUG nova.compute.manager [req-92c780b4-dfa9-4e9e-a60d-b03e1a53c40f req-60e4313d-74ae-48a6-86fd-2b51a16135a4 service nova] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Received event network-changed-fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 865.619372] env[62692]: DEBUG nova.compute.manager [req-92c780b4-dfa9-4e9e-a60d-b03e1a53c40f req-60e4313d-74ae-48a6-86fd-2b51a16135a4 service nova] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Refreshing instance network info cache due to event network-changed-fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 865.619617] env[62692]: DEBUG oslo_concurrency.lockutils [req-92c780b4-dfa9-4e9e-a60d-b03e1a53c40f req-60e4313d-74ae-48a6-86fd-2b51a16135a4 service nova] Acquiring lock "refresh_cache-c6a965d3-1c52-4f95-a226-9d15b7197ce7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.619768] env[62692]: DEBUG oslo_concurrency.lockutils [req-92c780b4-dfa9-4e9e-a60d-b03e1a53c40f req-60e4313d-74ae-48a6-86fd-2b51a16135a4 service nova] Acquired lock "refresh_cache-c6a965d3-1c52-4f95-a226-9d15b7197ce7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.619953] env[62692]: DEBUG nova.network.neutron [req-92c780b4-dfa9-4e9e-a60d-b03e1a53c40f req-60e4313d-74ae-48a6-86fd-2b51a16135a4 service nova] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Refreshing network info cache for port fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 865.741559] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141222, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.807359] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Releasing lock "refresh_cache-26c1ff5b-9eda-4131-a4ea-d4511a2364b7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.807359] env[62692]: DEBUG nova.compute.manager [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Instance network_info: |[{"id": "513b63b4-6498-4af7-b62f-9bdbc4f1e9ed", "address": "fa:16:3e:49:d0:ff", "network": {"id": "0e06e04a-5213-4d06-b3f1-b2b8f437e6d0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1533469056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "266c39fe585d41079d0278ecd1d1c6a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "678ebbe4-4c53-4eaf-a689-93981310f37d", "external-id": "nsx-vlan-transportzone-443", "segmentation_id": 443, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap513b63b4-64", "ovs_interfaceid": "513b63b4-6498-4af7-b62f-9bdbc4f1e9ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 865.808047] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:49:d0:ff', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '678ebbe4-4c53-4eaf-a689-93981310f37d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '513b63b4-6498-4af7-b62f-9bdbc4f1e9ed', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 865.816870] env[62692]: DEBUG oslo.service.loopingcall [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 865.817366] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 865.817791] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7ad3ac8b-781b-4129-8834-f8511a53acd0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.842869] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 865.842869] env[62692]: value = "task-1141223" [ 865.842869] env[62692]: _type = "Task" [ 865.842869] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.851446] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141223, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.921304] env[62692]: DEBUG nova.compute.manager [req-78c84a96-208f-44ae-b8dd-9d3968263c3a req-d43b7a2e-d3f2-45b8-9900-6b00175efa86 service nova] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Received event network-vif-plugged-513b63b4-6498-4af7-b62f-9bdbc4f1e9ed {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 865.921304] env[62692]: DEBUG oslo_concurrency.lockutils [req-78c84a96-208f-44ae-b8dd-9d3968263c3a req-d43b7a2e-d3f2-45b8-9900-6b00175efa86 service nova] Acquiring lock "26c1ff5b-9eda-4131-a4ea-d4511a2364b7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.922330] env[62692]: DEBUG oslo_concurrency.lockutils [req-78c84a96-208f-44ae-b8dd-9d3968263c3a req-d43b7a2e-d3f2-45b8-9900-6b00175efa86 service nova] Lock "26c1ff5b-9eda-4131-a4ea-d4511a2364b7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.922649] env[62692]: DEBUG oslo_concurrency.lockutils [req-78c84a96-208f-44ae-b8dd-9d3968263c3a req-d43b7a2e-d3f2-45b8-9900-6b00175efa86 service nova] Lock "26c1ff5b-9eda-4131-a4ea-d4511a2364b7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.922768] env[62692]: DEBUG nova.compute.manager [req-78c84a96-208f-44ae-b8dd-9d3968263c3a req-d43b7a2e-d3f2-45b8-9900-6b00175efa86 service nova] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] No waiting events found dispatching network-vif-plugged-513b63b4-6498-4af7-b62f-9bdbc4f1e9ed {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 865.923407] env[62692]: WARNING nova.compute.manager [req-78c84a96-208f-44ae-b8dd-9d3968263c3a req-d43b7a2e-d3f2-45b8-9900-6b00175efa86 service nova] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Received unexpected event network-vif-plugged-513b63b4-6498-4af7-b62f-9bdbc4f1e9ed for instance with vm_state building and task_state spawning. [ 865.923407] env[62692]: DEBUG nova.compute.manager [req-78c84a96-208f-44ae-b8dd-9d3968263c3a req-d43b7a2e-d3f2-45b8-9900-6b00175efa86 service nova] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Received event network-changed-513b63b4-6498-4af7-b62f-9bdbc4f1e9ed {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 865.923407] env[62692]: DEBUG nova.compute.manager [req-78c84a96-208f-44ae-b8dd-9d3968263c3a req-d43b7a2e-d3f2-45b8-9900-6b00175efa86 service nova] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Refreshing instance network info cache due to event network-changed-513b63b4-6498-4af7-b62f-9bdbc4f1e9ed. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 865.923596] env[62692]: DEBUG oslo_concurrency.lockutils [req-78c84a96-208f-44ae-b8dd-9d3968263c3a req-d43b7a2e-d3f2-45b8-9900-6b00175efa86 service nova] Acquiring lock "refresh_cache-26c1ff5b-9eda-4131-a4ea-d4511a2364b7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.924182] env[62692]: DEBUG oslo_concurrency.lockutils [req-78c84a96-208f-44ae-b8dd-9d3968263c3a req-d43b7a2e-d3f2-45b8-9900-6b00175efa86 service nova] Acquired lock "refresh_cache-26c1ff5b-9eda-4131-a4ea-d4511a2364b7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.924182] env[62692]: DEBUG nova.network.neutron [req-78c84a96-208f-44ae-b8dd-9d3968263c3a req-d43b7a2e-d3f2-45b8-9900-6b00175efa86 service nova] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Refreshing network info cache for port 513b63b4-6498-4af7-b62f-9bdbc4f1e9ed {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 865.958705] env[62692]: ERROR nova.scheduler.client.report [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [req-8060814c-fb8e-4994-a749-2185b9c2b6cd] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 470758a8-5efa-4875-aad5-f512a727752f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-8060814c-fb8e-4994-a749-2185b9c2b6cd"}]} [ 865.978152] env[62692]: DEBUG nova.scheduler.client.report [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Refreshing inventories for resource provider 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 865.996052] env[62692]: DEBUG nova.scheduler.client.report [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Updating ProviderTree inventory for provider 470758a8-5efa-4875-aad5-f512a727752f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 865.996297] env[62692]: DEBUG nova.compute.provider_tree [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 866.009729] env[62692]: DEBUG nova.scheduler.client.report [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Refreshing aggregate associations for resource provider 470758a8-5efa-4875-aad5-f512a727752f, aggregates: None {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 866.032705] env[62692]: DEBUG nova.scheduler.client.report [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Refreshing trait associations for resource provider 470758a8-5efa-4875-aad5-f512a727752f, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 866.050898] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a4950060-a599-4fa1-a856-e9bc0c94e03a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "refresh_cache-d99737d8-2eb0-40ee-b61e-6c736c84ea59" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.051139] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a4950060-a599-4fa1-a856-e9bc0c94e03a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired lock "refresh_cache-d99737d8-2eb0-40ee-b61e-6c736c84ea59" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.051332] env[62692]: DEBUG nova.network.neutron [None req-a4950060-a599-4fa1-a856-e9bc0c94e03a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 866.246623] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141222, 'name': ReconfigVM_Task, 'duration_secs': 0.822904} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.250174] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Reconfigured VM instance instance-0000003b to attach disk [datastore1] 754b3c65-1e4b-49d2-8980-095d975edb01/754b3c65-1e4b-49d2-8980-095d975edb01.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 866.250796] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8a12e555-eda9-4a5f-ad23-5149165ab421 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.259856] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 866.259856] env[62692]: value = "task-1141224" [ 866.259856] env[62692]: _type = "Task" [ 866.259856] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.273240] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141224, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.356290] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141223, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.469381] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac8f7ff1-0ab4-4fa4-8836-791f6e8c1a62 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.478504] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f374c81-4c74-486c-956c-80907a632e86 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.523047] env[62692]: DEBUG nova.compute.manager [req-d52767ff-cc8d-4958-92ed-b1492397f693 req-318e0040-e89f-468f-9c19-c563884063b5 service nova] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Received event network-changed-a291ddc4-8996-4755-b347-066f38e6ec7d {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 866.523262] env[62692]: DEBUG nova.compute.manager [req-d52767ff-cc8d-4958-92ed-b1492397f693 req-318e0040-e89f-468f-9c19-c563884063b5 service nova] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Refreshing instance network info cache due to event network-changed-a291ddc4-8996-4755-b347-066f38e6ec7d. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 866.523650] env[62692]: DEBUG oslo_concurrency.lockutils [req-d52767ff-cc8d-4958-92ed-b1492397f693 req-318e0040-e89f-468f-9c19-c563884063b5 service nova] Acquiring lock "refresh_cache-e2383a6a-3581-40fc-a0eb-6981acdbf54a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.523830] env[62692]: DEBUG oslo_concurrency.lockutils [req-d52767ff-cc8d-4958-92ed-b1492397f693 req-318e0040-e89f-468f-9c19-c563884063b5 service nova] Acquired lock "refresh_cache-e2383a6a-3581-40fc-a0eb-6981acdbf54a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.524018] env[62692]: DEBUG nova.network.neutron [req-d52767ff-cc8d-4958-92ed-b1492397f693 req-318e0040-e89f-468f-9c19-c563884063b5 service nova] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Refreshing network info cache for port a291ddc4-8996-4755-b347-066f38e6ec7d {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 866.526408] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7341ee2b-1583-41c4-b552-668a3a95b5e3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.538243] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d43c1db2-292b-43d8-8e61-949167238a76 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.564974] env[62692]: DEBUG nova.compute.provider_tree [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 866.669414] env[62692]: DEBUG nova.network.neutron [req-92c780b4-dfa9-4e9e-a60d-b03e1a53c40f req-60e4313d-74ae-48a6-86fd-2b51a16135a4 service nova] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Updated VIF entry in instance network info cache for port fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 866.669906] env[62692]: DEBUG nova.network.neutron [req-92c780b4-dfa9-4e9e-a60d-b03e1a53c40f req-60e4313d-74ae-48a6-86fd-2b51a16135a4 service nova] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Updating instance_info_cache with network_info: [{"id": "fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8", "address": "fa:16:3e:c7:7a:a0", "network": {"id": "04aacc3d-37b6-4a5c-ac4d-55313731e4f9", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1367536055-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef45ad1482dc4f5db51a5badc3cdce22", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a0734cc4-5718-45e2-9f98-0ded96880bef", "external-id": "nsx-vlan-transportzone-875", "segmentation_id": 875, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdd5ae79-fe", "ovs_interfaceid": "fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.773338] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141224, 'name': Rename_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.828710] env[62692]: DEBUG oslo_concurrency.lockutils [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Acquiring lock "18eb004f-a16b-477a-8d05-ed23a418b069" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.828710] env[62692]: DEBUG oslo_concurrency.lockutils [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Lock "18eb004f-a16b-477a-8d05-ed23a418b069" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.828710] env[62692]: DEBUG oslo_concurrency.lockutils [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Acquiring lock "18eb004f-a16b-477a-8d05-ed23a418b069-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.828710] env[62692]: DEBUG oslo_concurrency.lockutils [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Lock "18eb004f-a16b-477a-8d05-ed23a418b069-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.828710] env[62692]: DEBUG oslo_concurrency.lockutils [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Lock "18eb004f-a16b-477a-8d05-ed23a418b069-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.829777] env[62692]: INFO nova.compute.manager [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Terminating instance [ 866.836339] env[62692]: DEBUG nova.network.neutron [req-78c84a96-208f-44ae-b8dd-9d3968263c3a req-d43b7a2e-d3f2-45b8-9900-6b00175efa86 service nova] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Updated VIF entry in instance network info cache for port 513b63b4-6498-4af7-b62f-9bdbc4f1e9ed. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 866.836969] env[62692]: DEBUG nova.network.neutron [req-78c84a96-208f-44ae-b8dd-9d3968263c3a req-d43b7a2e-d3f2-45b8-9900-6b00175efa86 service nova] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Updating instance_info_cache with network_info: [{"id": "513b63b4-6498-4af7-b62f-9bdbc4f1e9ed", "address": "fa:16:3e:49:d0:ff", "network": {"id": "0e06e04a-5213-4d06-b3f1-b2b8f437e6d0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1533469056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "266c39fe585d41079d0278ecd1d1c6a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "678ebbe4-4c53-4eaf-a689-93981310f37d", "external-id": "nsx-vlan-transportzone-443", "segmentation_id": 443, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap513b63b4-64", "ovs_interfaceid": "513b63b4-6498-4af7-b62f-9bdbc4f1e9ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.839248] env[62692]: DEBUG nova.compute.manager [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 866.839608] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 866.843024] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5958a572-1ed1-4178-9452-b7390729c5cf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.857712] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 866.864896] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-793e3256-f848-4c58-a723-1f3b9fb72a58 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.867581] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141223, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.874412] env[62692]: DEBUG oslo_vmware.api [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Waiting for the task: (returnval){ [ 866.874412] env[62692]: value = "task-1141225" [ 866.874412] env[62692]: _type = "Task" [ 866.874412] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.885999] env[62692]: DEBUG oslo_vmware.api [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Task: {'id': task-1141225, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.925258] env[62692]: DEBUG oslo_concurrency.lockutils [None req-007d107e-2da6-4577-a382-c020ab02c70e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Acquiring lock "6b2d3385-a9af-432b-a1e7-c91adb2e1f69" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.925600] env[62692]: DEBUG oslo_concurrency.lockutils [None req-007d107e-2da6-4577-a382-c020ab02c70e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Lock "6b2d3385-a9af-432b-a1e7-c91adb2e1f69" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.925860] env[62692]: INFO nova.compute.manager [None req-007d107e-2da6-4577-a382-c020ab02c70e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Rebooting instance [ 867.020886] env[62692]: DEBUG nova.network.neutron [None req-a4950060-a599-4fa1-a856-e9bc0c94e03a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Updating instance_info_cache with network_info: [{"id": "7f2cabdf-cbda-4cf7-a518-13f619716d12", "address": "fa:16:3e:6b:22:0c", "network": {"id": "f70727d6-a9aa-44b0-a244-2533b3f96571", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-284105837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1037e5d436bd429391159b87a712f6e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f2cabdf-cb", "ovs_interfaceid": "7f2cabdf-cbda-4cf7-a518-13f619716d12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.109289] env[62692]: DEBUG nova.scheduler.client.report [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Updated inventory for provider 470758a8-5efa-4875-aad5-f512a727752f with generation 78 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 867.109289] env[62692]: DEBUG nova.compute.provider_tree [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Updating resource provider 470758a8-5efa-4875-aad5-f512a727752f generation from 78 to 79 during operation: update_inventory {{(pid=62692) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 867.109289] env[62692]: DEBUG nova.compute.provider_tree [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 867.172759] env[62692]: DEBUG oslo_concurrency.lockutils [req-92c780b4-dfa9-4e9e-a60d-b03e1a53c40f req-60e4313d-74ae-48a6-86fd-2b51a16135a4 service nova] Releasing lock "refresh_cache-c6a965d3-1c52-4f95-a226-9d15b7197ce7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.272325] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141224, 'name': Rename_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.346592] env[62692]: DEBUG oslo_concurrency.lockutils [req-78c84a96-208f-44ae-b8dd-9d3968263c3a req-d43b7a2e-d3f2-45b8-9900-6b00175efa86 service nova] Releasing lock "refresh_cache-26c1ff5b-9eda-4131-a4ea-d4511a2364b7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.356477] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141223, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.357508] env[62692]: DEBUG nova.network.neutron [req-d52767ff-cc8d-4958-92ed-b1492397f693 req-318e0040-e89f-468f-9c19-c563884063b5 service nova] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Updated VIF entry in instance network info cache for port a291ddc4-8996-4755-b347-066f38e6ec7d. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 867.357854] env[62692]: DEBUG nova.network.neutron [req-d52767ff-cc8d-4958-92ed-b1492397f693 req-318e0040-e89f-468f-9c19-c563884063b5 service nova] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Updating instance_info_cache with network_info: [{"id": "a291ddc4-8996-4755-b347-066f38e6ec7d", "address": "fa:16:3e:7b:10:2a", "network": {"id": "2febb660-329d-403c-8c2c-f11c0f5f547c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1806650357-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37921f67664544478cb2568def068b11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa291ddc4-89", "ovs_interfaceid": "a291ddc4-8996-4755-b347-066f38e6ec7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.385593] env[62692]: DEBUG oslo_vmware.api [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Task: {'id': task-1141225, 'name': PowerOffVM_Task, 'duration_secs': 0.328865} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.385885] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 867.386075] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 867.386347] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-802ec5d8-cc92-44dd-8f78-546b84416eb4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.458431] env[62692]: DEBUG oslo_concurrency.lockutils [None req-007d107e-2da6-4577-a382-c020ab02c70e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Acquiring lock "refresh_cache-6b2d3385-a9af-432b-a1e7-c91adb2e1f69" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.458680] env[62692]: DEBUG oslo_concurrency.lockutils [None req-007d107e-2da6-4577-a382-c020ab02c70e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Acquired lock "refresh_cache-6b2d3385-a9af-432b-a1e7-c91adb2e1f69" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.458822] env[62692]: DEBUG nova.network.neutron [None req-007d107e-2da6-4577-a382-c020ab02c70e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 867.472298] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 867.472586] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 867.472807] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Deleting the datastore file [datastore2] 18eb004f-a16b-477a-8d05-ed23a418b069 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 867.473687] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-be7b8d65-aa97-468d-ae29-56ff7987e23d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.482315] env[62692]: DEBUG oslo_vmware.api [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Waiting for the task: (returnval){ [ 867.482315] env[62692]: value = "task-1141227" [ 867.482315] env[62692]: _type = "Task" [ 867.482315] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.492404] env[62692]: DEBUG oslo_vmware.api [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Task: {'id': task-1141227, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.493963] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Acquiring lock "c6a965d3-1c52-4f95-a226-9d15b7197ce7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.494240] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Lock "c6a965d3-1c52-4f95-a226-9d15b7197ce7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.494480] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Acquiring lock "c6a965d3-1c52-4f95-a226-9d15b7197ce7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.494675] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Lock "c6a965d3-1c52-4f95-a226-9d15b7197ce7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.494865] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Lock "c6a965d3-1c52-4f95-a226-9d15b7197ce7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.497305] env[62692]: INFO nova.compute.manager [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Terminating instance [ 867.499593] env[62692]: DEBUG nova.compute.manager [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 867.499800] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 867.500626] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84dc5ac5-419f-405d-8aa8-09b82dbba8e5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.510768] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 867.511037] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-504d5512-ac7b-4979-8206-fcd70aa800ef {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.520110] env[62692]: DEBUG oslo_vmware.api [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Waiting for the task: (returnval){ [ 867.520110] env[62692]: value = "task-1141228" [ 867.520110] env[62692]: _type = "Task" [ 867.520110] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.524785] env[62692]: DEBUG oslo_concurrency.lockutils [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.525384] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a4950060-a599-4fa1-a856-e9bc0c94e03a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Releasing lock "refresh_cache-d99737d8-2eb0-40ee-b61e-6c736c84ea59" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.528437] env[62692]: DEBUG nova.compute.manager [None req-a4950060-a599-4fa1-a856-e9bc0c94e03a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 867.529546] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-732a0b75-2d83-42c5-9ce2-a649fc4498e3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.536058] env[62692]: DEBUG oslo_vmware.api [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Task: {'id': task-1141228, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.616027] env[62692]: DEBUG oslo_concurrency.lockutils [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.299s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.619094] env[62692]: DEBUG oslo_concurrency.lockutils [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.840s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.621937] env[62692]: INFO nova.compute.claims [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 867.743853] env[62692]: INFO nova.scheduler.client.report [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Deleted allocations for instance b1d6409a-2733-470f-a929-672fe1631b1b [ 867.773909] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141224, 'name': Rename_Task, 'duration_secs': 1.190578} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.774414] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 867.776078] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eaf0a117-39db-487d-8046-5c2fd2f84ad3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.785049] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 867.785049] env[62692]: value = "task-1141229" [ 867.785049] env[62692]: _type = "Task" [ 867.785049] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.796249] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "interface-27583cb4-0478-44d8-8970-478682271513-1b690180-b6c8-4a42-96de-cfd6f25d1791" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.796668] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "interface-27583cb4-0478-44d8-8970-478682271513-1b690180-b6c8-4a42-96de-cfd6f25d1791" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.811143] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141229, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.859849] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141223, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.860549] env[62692]: DEBUG oslo_concurrency.lockutils [req-d52767ff-cc8d-4958-92ed-b1492397f693 req-318e0040-e89f-468f-9c19-c563884063b5 service nova] Releasing lock "refresh_cache-e2383a6a-3581-40fc-a0eb-6981acdbf54a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.996116] env[62692]: DEBUG oslo_vmware.api [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Task: {'id': task-1141227, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.247161} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.996577] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 867.996886] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 867.997409] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 867.997861] env[62692]: INFO nova.compute.manager [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Took 1.16 seconds to destroy the instance on the hypervisor. [ 867.998310] env[62692]: DEBUG oslo.service.loopingcall [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.998687] env[62692]: DEBUG nova.compute.manager [-] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 867.998991] env[62692]: DEBUG nova.network.neutron [-] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 868.032026] env[62692]: DEBUG oslo_vmware.api [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Task: {'id': task-1141228, 'name': PowerOffVM_Task, 'duration_secs': 0.259277} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.032557] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 868.032831] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 868.034871] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-de2724d0-c4b7-417f-85ba-0a3c3d747391 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.037742] env[62692]: DEBUG nova.compute.manager [req-02132a33-a1fa-4f43-85c2-4892041a73a5 req-e6594744-15fb-4217-b4d9-7ff10617ec25 service nova] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Received event network-changed-318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 868.038012] env[62692]: DEBUG nova.compute.manager [req-02132a33-a1fa-4f43-85c2-4892041a73a5 req-e6594744-15fb-4217-b4d9-7ff10617ec25 service nova] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Refreshing instance network info cache due to event network-changed-318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 868.038277] env[62692]: DEBUG oslo_concurrency.lockutils [req-02132a33-a1fa-4f43-85c2-4892041a73a5 req-e6594744-15fb-4217-b4d9-7ff10617ec25 service nova] Acquiring lock "refresh_cache-6b2d3385-a9af-432b-a1e7-c91adb2e1f69" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.114192] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 868.114192] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 868.114192] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Deleting the datastore file [datastore2] c6a965d3-1c52-4f95-a226-9d15b7197ce7 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 868.114192] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ddfdf442-9723-4702-b153-d7cbe573a45c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.131156] env[62692]: DEBUG oslo_vmware.api [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Waiting for the task: (returnval){ [ 868.131156] env[62692]: value = "task-1141231" [ 868.131156] env[62692]: _type = "Task" [ 868.131156] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.145278] env[62692]: DEBUG oslo_vmware.api [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Task: {'id': task-1141231, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.213021] env[62692]: DEBUG oslo_concurrency.lockutils [None req-fd1fe936-78a2-4264-83e7-fd8f51817c1e tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Acquiring lock "81ce340c-fbef-4932-983c-595843530dbc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.213021] env[62692]: DEBUG oslo_concurrency.lockutils [None req-fd1fe936-78a2-4264-83e7-fd8f51817c1e tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Lock "81ce340c-fbef-4932-983c-595843530dbc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.213021] env[62692]: DEBUG oslo_concurrency.lockutils [None req-fd1fe936-78a2-4264-83e7-fd8f51817c1e tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Acquiring lock "81ce340c-fbef-4932-983c-595843530dbc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.213021] env[62692]: DEBUG oslo_concurrency.lockutils [None req-fd1fe936-78a2-4264-83e7-fd8f51817c1e tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Lock "81ce340c-fbef-4932-983c-595843530dbc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.213021] env[62692]: DEBUG oslo_concurrency.lockutils [None req-fd1fe936-78a2-4264-83e7-fd8f51817c1e tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Lock "81ce340c-fbef-4932-983c-595843530dbc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.215952] env[62692]: INFO nova.compute.manager [None req-fd1fe936-78a2-4264-83e7-fd8f51817c1e tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Terminating instance [ 868.219802] env[62692]: DEBUG nova.compute.manager [None req-fd1fe936-78a2-4264-83e7-fd8f51817c1e tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 868.223020] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f9077bd7-6c31-4ea3-97f8-b4300c3ecb41 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.235899] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3d78165-66c6-467c-9e13-263701bf01df {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.260947] env[62692]: DEBUG oslo_concurrency.lockutils [None req-039271c3-5061-4b13-8ba4-5ee31b38bebb tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "b1d6409a-2733-470f-a929-672fe1631b1b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.093s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.290899] env[62692]: WARNING nova.virt.vmwareapi.driver [None req-fd1fe936-78a2-4264-83e7-fd8f51817c1e tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 81ce340c-fbef-4932-983c-595843530dbc could not be found. [ 868.291212] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-fd1fe936-78a2-4264-83e7-fd8f51817c1e tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 868.295640] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-69c6935f-f8f2-40f3-8dbd-1b84f603bc98 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.299319] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "27583cb4-0478-44d8-8970-478682271513" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.299460] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "27583cb4-0478-44d8-8970-478682271513" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.300424] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a83d7950-23c6-496c-8649-5688d0b7eade {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.307664] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141229, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.311122] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec9cf58d-792a-4599-867a-420a9d2836d5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.343036] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-323378ba-161c-4a3d-89cb-dd6f4a15232d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.381528] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Reconfiguring VM to detach interface {{(pid=62692) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 868.402555] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-124d8867-c6f3-4c43-a51e-376fb065796c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.427494] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-fd1fe936-78a2-4264-83e7-fd8f51817c1e tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 81ce340c-fbef-4932-983c-595843530dbc could not be found. [ 868.427729] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-fd1fe936-78a2-4264-83e7-fd8f51817c1e tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 868.427848] env[62692]: INFO nova.compute.manager [None req-fd1fe936-78a2-4264-83e7-fd8f51817c1e tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Took 0.21 seconds to destroy the instance on the hypervisor. [ 868.428253] env[62692]: DEBUG oslo.service.loopingcall [None req-fd1fe936-78a2-4264-83e7-fd8f51817c1e tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 868.436077] env[62692]: DEBUG nova.compute.manager [-] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 868.436375] env[62692]: DEBUG nova.network.neutron [-] [instance: 81ce340c-fbef-4932-983c-595843530dbc] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 868.438304] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141223, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.444921] env[62692]: DEBUG oslo_vmware.api [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 868.444921] env[62692]: value = "task-1141232" [ 868.444921] env[62692]: _type = "Task" [ 868.444921] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.455378] env[62692]: DEBUG oslo_vmware.api [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141232, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.531020] env[62692]: DEBUG nova.network.neutron [None req-007d107e-2da6-4577-a382-c020ab02c70e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Updating instance_info_cache with network_info: [{"id": "318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db", "address": "fa:16:3e:32:9a:8d", "network": {"id": "7e7a1dcd-36ba-4a25-affa-6bd7fe895ade", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-508011613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9667b0b7abaa4bf0a7e54e9d540df728", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap318dc9c9-0a", "ovs_interfaceid": "318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.552082] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-269771db-32c8-40da-a095-fbd91123c0de {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.562555] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a4950060-a599-4fa1-a856-e9bc0c94e03a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Doing hard reboot of VM {{(pid=62692) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 868.566162] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-462bad6e-6a96-4b31-8630-7ac7dc5cb232 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.574386] env[62692]: DEBUG oslo_vmware.api [None req-a4950060-a599-4fa1-a856-e9bc0c94e03a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 868.574386] env[62692]: value = "task-1141233" [ 868.574386] env[62692]: _type = "Task" [ 868.574386] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.587214] env[62692]: DEBUG oslo_vmware.api [None req-a4950060-a599-4fa1-a856-e9bc0c94e03a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141233, 'name': ResetVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.647809] env[62692]: DEBUG oslo_vmware.api [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Task: {'id': task-1141231, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.326841} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.648144] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 868.648285] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 868.648466] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 868.648643] env[62692]: INFO nova.compute.manager [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Took 1.15 seconds to destroy the instance on the hypervisor. [ 868.648904] env[62692]: DEBUG oslo.service.loopingcall [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 868.649156] env[62692]: DEBUG nova.compute.manager [-] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 868.649256] env[62692]: DEBUG nova.network.neutron [-] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 868.808275] env[62692]: DEBUG oslo_vmware.api [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141229, 'name': PowerOnVM_Task, 'duration_secs': 0.759562} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.808552] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 868.808924] env[62692]: INFO nova.compute.manager [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Took 10.72 seconds to spawn the instance on the hypervisor. [ 868.809532] env[62692]: DEBUG nova.compute.manager [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 868.810688] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75844cd2-2e4c-4ecc-b852-5157243759af {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.863673] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141223, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.959701] env[62692]: DEBUG oslo_vmware.api [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141232, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.002794] env[62692]: DEBUG nova.network.neutron [-] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.032703] env[62692]: DEBUG oslo_concurrency.lockutils [None req-007d107e-2da6-4577-a382-c020ab02c70e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Releasing lock "refresh_cache-6b2d3385-a9af-432b-a1e7-c91adb2e1f69" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.035199] env[62692]: DEBUG oslo_concurrency.lockutils [req-02132a33-a1fa-4f43-85c2-4892041a73a5 req-e6594744-15fb-4217-b4d9-7ff10617ec25 service nova] Acquired lock "refresh_cache-6b2d3385-a9af-432b-a1e7-c91adb2e1f69" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.035596] env[62692]: DEBUG nova.network.neutron [req-02132a33-a1fa-4f43-85c2-4892041a73a5 req-e6594744-15fb-4217-b4d9-7ff10617ec25 service nova] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Refreshing network info cache for port 318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 869.038988] env[62692]: DEBUG nova.compute.manager [None req-007d107e-2da6-4577-a382-c020ab02c70e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 869.040026] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4266e2e-8624-4d72-a5ed-f62f5478d610 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.099570] env[62692]: DEBUG oslo_vmware.api [None req-a4950060-a599-4fa1-a856-e9bc0c94e03a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141233, 'name': ResetVM_Task, 'duration_secs': 0.109587} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.100524] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a4950060-a599-4fa1-a856-e9bc0c94e03a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Did hard reboot of VM {{(pid=62692) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 869.100715] env[62692]: DEBUG nova.compute.manager [None req-a4950060-a599-4fa1-a856-e9bc0c94e03a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 869.101531] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c92907e-1eb9-4e76-adf4-933cd62ee1c6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.170527] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26b0430c-3d70-403c-a4e5-b0fc986716df {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.180654] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62399557-acff-4a1f-8674-bbc6303cca39 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.221016] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39a1b8ed-da05-4b6e-870f-05e0e991c06f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.231576] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78d625cb-02a2-4abe-a620-afb4b04b39ea {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.251642] env[62692]: DEBUG nova.compute.provider_tree [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 869.336586] env[62692]: INFO nova.compute.manager [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Took 37.85 seconds to build instance. [ 869.364426] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141223, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.456596] env[62692]: DEBUG oslo_vmware.api [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141232, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.505581] env[62692]: INFO nova.compute.manager [-] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Took 1.51 seconds to deallocate network for instance. [ 869.593358] env[62692]: DEBUG nova.network.neutron [-] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.625854] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a4950060-a599-4fa1-a856-e9bc0c94e03a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "d99737d8-2eb0-40ee-b61e-6c736c84ea59" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.101s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.775553] env[62692]: ERROR nova.scheduler.client.report [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [req-8bdb62b3-0843-46da-8fa1-e9d70ddc0c09] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 470758a8-5efa-4875-aad5-f512a727752f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-8bdb62b3-0843-46da-8fa1-e9d70ddc0c09"}]} [ 869.807049] env[62692]: DEBUG nova.scheduler.client.report [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Refreshing inventories for resource provider 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 869.824567] env[62692]: DEBUG nova.scheduler.client.report [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Updating ProviderTree inventory for provider 470758a8-5efa-4875-aad5-f512a727752f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 869.825208] env[62692]: DEBUG nova.compute.provider_tree [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 869.838774] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f1180dcc-6c8a-41be-8de3-5507fd639142 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "754b3c65-1e4b-49d2-8980-095d975edb01" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.081s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.842848] env[62692]: DEBUG nova.scheduler.client.report [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Refreshing aggregate associations for resource provider 470758a8-5efa-4875-aad5-f512a727752f, aggregates: None {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 869.868291] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141223, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.869425] env[62692]: DEBUG nova.scheduler.client.report [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Refreshing trait associations for resource provider 470758a8-5efa-4875-aad5-f512a727752f, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 869.941232] env[62692]: DEBUG nova.network.neutron [req-02132a33-a1fa-4f43-85c2-4892041a73a5 req-e6594744-15fb-4217-b4d9-7ff10617ec25 service nova] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Updated VIF entry in instance network info cache for port 318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 869.941232] env[62692]: DEBUG nova.network.neutron [req-02132a33-a1fa-4f43-85c2-4892041a73a5 req-e6594744-15fb-4217-b4d9-7ff10617ec25 service nova] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Updating instance_info_cache with network_info: [{"id": "318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db", "address": "fa:16:3e:32:9a:8d", "network": {"id": "7e7a1dcd-36ba-4a25-affa-6bd7fe895ade", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-508011613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9667b0b7abaa4bf0a7e54e9d540df728", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap318dc9c9-0a", "ovs_interfaceid": "318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.963673] env[62692]: DEBUG oslo_vmware.api [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141232, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.020099] env[62692]: DEBUG oslo_concurrency.lockutils [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.069064] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d902610-862c-4de3-88ba-80e2297940b6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.083659] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-007d107e-2da6-4577-a382-c020ab02c70e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Doing hard reboot of VM {{(pid=62692) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 870.083659] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-ddf9558a-5400-41e9-aeb0-7effeadab288 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.096542] env[62692]: INFO nova.compute.manager [-] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Took 1.45 seconds to deallocate network for instance. [ 870.101173] env[62692]: DEBUG oslo_vmware.api [None req-007d107e-2da6-4577-a382-c020ab02c70e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Waiting for the task: (returnval){ [ 870.101173] env[62692]: value = "task-1141234" [ 870.101173] env[62692]: _type = "Task" [ 870.101173] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.118472] env[62692]: DEBUG oslo_vmware.api [None req-007d107e-2da6-4577-a382-c020ab02c70e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1141234, 'name': ResetVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.121623] env[62692]: DEBUG oslo_concurrency.lockutils [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "3fd2959b-80cb-470c-8c2a-40b7630458ed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.121861] env[62692]: DEBUG oslo_concurrency.lockutils [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "3fd2959b-80cb-470c-8c2a-40b7630458ed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.166270] env[62692]: DEBUG nova.compute.manager [req-15264e77-3d65-49c3-aac8-e8fa7c02f490 req-afac48c0-11ec-4d41-8ad0-8be82668679f service nova] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Received event network-vif-deleted-6b631aba-53ff-466f-b003-cc92a2b87144 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 870.166270] env[62692]: INFO nova.compute.manager [req-15264e77-3d65-49c3-aac8-e8fa7c02f490 req-afac48c0-11ec-4d41-8ad0-8be82668679f service nova] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Neutron deleted interface 6b631aba-53ff-466f-b003-cc92a2b87144; detaching it from the instance and deleting it from the info cache [ 870.166696] env[62692]: DEBUG nova.network.neutron [req-15264e77-3d65-49c3-aac8-e8fa7c02f490 req-afac48c0-11ec-4d41-8ad0-8be82668679f service nova] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.284020] env[62692]: DEBUG nova.network.neutron [-] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.369137] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141223, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.374380] env[62692]: DEBUG nova.compute.manager [req-6a87c125-bb31-4679-85dc-eb756bf632e4 req-81860647-1028-4b94-9d35-e87c67bd99a0 service nova] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Received event network-vif-deleted-ddc1a77b-94d9-468b-ab57-8aef5343e580 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 870.374602] env[62692]: DEBUG nova.compute.manager [req-6a87c125-bb31-4679-85dc-eb756bf632e4 req-81860647-1028-4b94-9d35-e87c67bd99a0 service nova] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Received event network-vif-deleted-fdd5ae79-fe9d-4ed5-9f80-4811d8097ce8 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 870.434625] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9956eb95-02be-4387-9673-61cdc3e33b42 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.443655] env[62692]: DEBUG oslo_concurrency.lockutils [req-02132a33-a1fa-4f43-85c2-4892041a73a5 req-e6594744-15fb-4217-b4d9-7ff10617ec25 service nova] Releasing lock "refresh_cache-6b2d3385-a9af-432b-a1e7-c91adb2e1f69" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.445409] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01bf0474-2ccf-4dfa-886b-a9a06a4a1034 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.461132] env[62692]: DEBUG oslo_vmware.api [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141232, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.488267] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51b2e22b-e983-40e1-a033-ed0b4b38836a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.497737] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a436a91-f8c6-4f51-978c-18bb393269c0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.513786] env[62692]: DEBUG nova.compute.provider_tree [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 870.608062] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.619740] env[62692]: DEBUG oslo_vmware.api [None req-007d107e-2da6-4577-a382-c020ab02c70e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1141234, 'name': ResetVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.624577] env[62692]: DEBUG nova.compute.manager [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 870.671201] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-32d45fd7-d404-4ffb-bfc9-80753a0c7982 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.682764] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-621112e0-a5ac-4e98-90ff-17ee0c4e5d6b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.725360] env[62692]: DEBUG nova.compute.manager [req-15264e77-3d65-49c3-aac8-e8fa7c02f490 req-afac48c0-11ec-4d41-8ad0-8be82668679f service nova] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Detach interface failed, port_id=6b631aba-53ff-466f-b003-cc92a2b87144, reason: Instance 81ce340c-fbef-4932-983c-595843530dbc could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 870.790179] env[62692]: INFO nova.compute.manager [-] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Took 2.35 seconds to deallocate network for instance. [ 870.867689] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141223, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.962037] env[62692]: DEBUG oslo_vmware.api [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141232, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.054692] env[62692]: DEBUG nova.scheduler.client.report [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Updated inventory for provider 470758a8-5efa-4875-aad5-f512a727752f with generation 80 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 871.054978] env[62692]: DEBUG nova.compute.provider_tree [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Updating resource provider 470758a8-5efa-4875-aad5-f512a727752f generation from 80 to 81 during operation: update_inventory {{(pid=62692) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 871.055181] env[62692]: DEBUG nova.compute.provider_tree [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 871.117305] env[62692]: DEBUG oslo_vmware.api [None req-007d107e-2da6-4577-a382-c020ab02c70e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1141234, 'name': ResetVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.151872] env[62692]: DEBUG oslo_concurrency.lockutils [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.343297] env[62692]: INFO nova.compute.manager [None req-fd1fe936-78a2-4264-83e7-fd8f51817c1e tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Took 0.55 seconds to detach 1 volumes for instance. [ 871.346028] env[62692]: DEBUG nova.compute.manager [None req-fd1fe936-78a2-4264-83e7-fd8f51817c1e tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Deleting volume: 9886d0c7-5df2-4bce-8853-faa1431b1865 {{(pid=62692) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 871.369745] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141223, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.463241] env[62692]: DEBUG oslo_vmware.api [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141232, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.552805] env[62692]: DEBUG oslo_vmware.rw_handles [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d91c56-689f-1ca1-03bd-93bd6a3add50/disk-0.vmdk. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 871.553930] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8fd8ddf-6367-4e3b-989d-74384df6642b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.561874] env[62692]: DEBUG oslo_concurrency.lockutils [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.943s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.562273] env[62692]: DEBUG nova.compute.manager [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 871.565563] env[62692]: DEBUG oslo_vmware.rw_handles [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d91c56-689f-1ca1-03bd-93bd6a3add50/disk-0.vmdk is in state: ready. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 871.565752] env[62692]: ERROR oslo_vmware.rw_handles [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d91c56-689f-1ca1-03bd-93bd6a3add50/disk-0.vmdk due to incomplete transfer. [ 871.566276] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.238s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.566503] env[62692]: DEBUG nova.objects.instance [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Lazy-loading 'resources' on Instance uuid ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 871.567780] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-74a9bf08-c145-4e16-a724-7ba2e8700108 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.576558] env[62692]: DEBUG oslo_vmware.rw_handles [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d91c56-689f-1ca1-03bd-93bd6a3add50/disk-0.vmdk. {{(pid=62692) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 871.576787] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Uploaded image 427c2d72-be26-42e4-9f8c-ca628806da41 to the Glance image server {{(pid=62692) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 871.579430] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Destroying the VM {{(pid=62692) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 871.580048] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-391228b5-7f0f-4110-81a1-1a8df28b0d2d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.588308] env[62692]: DEBUG oslo_vmware.api [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 871.588308] env[62692]: value = "task-1141236" [ 871.588308] env[62692]: _type = "Task" [ 871.588308] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.597703] env[62692]: DEBUG oslo_vmware.api [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141236, 'name': Destroy_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.618339] env[62692]: DEBUG oslo_vmware.api [None req-007d107e-2da6-4577-a382-c020ab02c70e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1141234, 'name': ResetVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.869414] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141223, 'name': CreateVM_Task, 'duration_secs': 5.840199} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.869635] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 871.870799] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.872040] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.872977] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 871.873350] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9bcebc38-dd70-4e25-b520-30020c788dfd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.884624] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 871.884624] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]526fac33-9672-6365-1a43-23e159ece93d" [ 871.884624] env[62692]: _type = "Task" [ 871.884624] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.895736] env[62692]: DEBUG oslo_concurrency.lockutils [None req-fd1fe936-78a2-4264-83e7-fd8f51817c1e tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.896138] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]526fac33-9672-6365-1a43-23e159ece93d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.963390] env[62692]: DEBUG oslo_vmware.api [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141232, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.068803] env[62692]: DEBUG nova.compute.utils [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 872.070316] env[62692]: DEBUG nova.compute.manager [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 872.070497] env[62692]: DEBUG nova.network.neutron [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 872.102868] env[62692]: DEBUG oslo_vmware.api [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141236, 'name': Destroy_Task} progress is 33%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.122023] env[62692]: DEBUG oslo_vmware.api [None req-007d107e-2da6-4577-a382-c020ab02c70e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1141234, 'name': ResetVM_Task, 'duration_secs': 1.989744} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.122988] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-007d107e-2da6-4577-a382-c020ab02c70e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Did hard reboot of VM {{(pid=62692) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 872.123886] env[62692]: DEBUG nova.compute.manager [None req-007d107e-2da6-4577-a382-c020ab02c70e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 872.125886] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19f9d98f-6351-4cd3-9e64-745f38aa01a1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.151854] env[62692]: DEBUG nova.policy [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9eb93376fdfe4dcc97a6220b076fd845', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f7abef2aee3c4669801439f8e0fdad22', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 872.403070] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]526fac33-9672-6365-1a43-23e159ece93d, 'name': SearchDatastore_Task, 'duration_secs': 0.139119} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.407021] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.407021] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 872.407021] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.407021] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.407021] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 872.407021] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8dd5b249-a017-4614-b208-34d89b345b94 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.418849] env[62692]: DEBUG nova.compute.manager [req-254e5b7d-4020-42b1-802f-330024c8db96 req-1663b8b3-6905-44eb-9bb3-c3f70874a325 service nova] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Received event network-changed-152b902d-a99b-44d9-974f-71348e613fc9 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 872.419131] env[62692]: DEBUG nova.compute.manager [req-254e5b7d-4020-42b1-802f-330024c8db96 req-1663b8b3-6905-44eb-9bb3-c3f70874a325 service nova] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Refreshing instance network info cache due to event network-changed-152b902d-a99b-44d9-974f-71348e613fc9. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 872.419351] env[62692]: DEBUG oslo_concurrency.lockutils [req-254e5b7d-4020-42b1-802f-330024c8db96 req-1663b8b3-6905-44eb-9bb3-c3f70874a325 service nova] Acquiring lock "refresh_cache-754b3c65-1e4b-49d2-8980-095d975edb01" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.419494] env[62692]: DEBUG oslo_concurrency.lockutils [req-254e5b7d-4020-42b1-802f-330024c8db96 req-1663b8b3-6905-44eb-9bb3-c3f70874a325 service nova] Acquired lock "refresh_cache-754b3c65-1e4b-49d2-8980-095d975edb01" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.419655] env[62692]: DEBUG nova.network.neutron [req-254e5b7d-4020-42b1-802f-330024c8db96 req-1663b8b3-6905-44eb-9bb3-c3f70874a325 service nova] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Refreshing network info cache for port 152b902d-a99b-44d9-974f-71348e613fc9 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 872.422528] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 872.422726] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 872.423671] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c76f0308-d9b2-451b-99e6-c3e51dda6fc2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.430725] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 872.430725] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]529998c3-fc2d-da00-3446-13e9f33452ac" [ 872.430725] env[62692]: _type = "Task" [ 872.430725] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.440337] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]529998c3-fc2d-da00-3446-13e9f33452ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.467715] env[62692]: DEBUG oslo_vmware.api [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141232, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.525353] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b43943fb-54d9-4e8a-a045-794ca9b8618f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.534223] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7131758-56ff-40dc-b3e3-cc75b422bd77 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.578847] env[62692]: DEBUG nova.compute.manager [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 872.584677] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f87febd-e846-49f8-80ef-edfa0a502645 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.589863] env[62692]: DEBUG nova.network.neutron [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Successfully created port: 49431714-cc36-41a1-a843-809653e38ccc {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 872.606619] env[62692]: DEBUG oslo_vmware.api [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141236, 'name': Destroy_Task} progress is 33%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.608073] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3802245a-59f0-4a0c-b9bc-56dbb599e829 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.625018] env[62692]: DEBUG nova.compute.provider_tree [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 872.642046] env[62692]: DEBUG oslo_concurrency.lockutils [None req-007d107e-2da6-4577-a382-c020ab02c70e tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Lock "6b2d3385-a9af-432b-a1e7-c91adb2e1f69" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 5.716s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.945845] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]529998c3-fc2d-da00-3446-13e9f33452ac, 'name': SearchDatastore_Task, 'duration_secs': 0.036273} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.946688] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba3a700a-3380-4f04-a739-b1b7d71f5bc4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.959313] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 872.959313] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5274db67-5a8c-ecab-4054-ef2b7987ecb8" [ 872.959313] env[62692]: _type = "Task" [ 872.959313] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.967203] env[62692]: DEBUG oslo_vmware.api [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141232, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.972570] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5274db67-5a8c-ecab-4054-ef2b7987ecb8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.092366] env[62692]: INFO nova.virt.block_device [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Booting with volume 4e85e9f0-749c-4394-8d51-fe35fa42cd6d at /dev/sda [ 873.107836] env[62692]: DEBUG oslo_vmware.api [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141236, 'name': Destroy_Task, 'duration_secs': 1.067879} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.115042] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Destroyed the VM [ 873.116077] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Deleting Snapshot of the VM instance {{(pid=62692) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 873.118642] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-aa24600e-bf48-4070-8686-a2d2e6294cc7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.128403] env[62692]: DEBUG nova.scheduler.client.report [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 873.134983] env[62692]: DEBUG oslo_vmware.api [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 873.134983] env[62692]: value = "task-1141237" [ 873.134983] env[62692]: _type = "Task" [ 873.134983] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.144231] env[62692]: DEBUG oslo_vmware.api [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141237, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.154284] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-86fbf596-0d8e-402e-be98-34bf6e0b3ea6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.167160] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0bc4f96-b7c9-4034-9564-6f9247c3d370 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.214696] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ac50007e-a70a-4275-a7a7-cb80fde5d242 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.226172] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c3d869a-ed61-43fd-82fd-c7cf93ad2e77 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.265653] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87dac3ff-b407-4537-8ed8-f32d12b6df34 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.273952] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0b0bc61-f6c5-49c5-84c1-b11bbc2f888a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.293493] env[62692]: DEBUG nova.virt.block_device [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Updating existing volume attachment record: d157af8f-5cb0-47ae-9b47-5742bdf3c1a5 {{(pid=62692) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 873.381202] env[62692]: DEBUG nova.network.neutron [req-254e5b7d-4020-42b1-802f-330024c8db96 req-1663b8b3-6905-44eb-9bb3-c3f70874a325 service nova] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Updated VIF entry in instance network info cache for port 152b902d-a99b-44d9-974f-71348e613fc9. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 873.381382] env[62692]: DEBUG nova.network.neutron [req-254e5b7d-4020-42b1-802f-330024c8db96 req-1663b8b3-6905-44eb-9bb3-c3f70874a325 service nova] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Updating instance_info_cache with network_info: [{"id": "152b902d-a99b-44d9-974f-71348e613fc9", "address": "fa:16:3e:11:96:8e", "network": {"id": "e9c538a3-7723-42cb-85b7-5a1597e49cef", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1878627765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.221", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71b00c566cea4b21811ea1e1ef5ba0f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap152b902d-a9", "ovs_interfaceid": "152b902d-a99b-44d9-974f-71348e613fc9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.417367] env[62692]: DEBUG oslo_concurrency.lockutils [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "44699df6-5262-4700-a096-8c0e450323d2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.417367] env[62692]: DEBUG oslo_concurrency.lockutils [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "44699df6-5262-4700-a096-8c0e450323d2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.468949] env[62692]: DEBUG oslo_vmware.api [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141232, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.476538] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5274db67-5a8c-ecab-4054-ef2b7987ecb8, 'name': SearchDatastore_Task, 'duration_secs': 0.067555} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.476863] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.477310] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 26c1ff5b-9eda-4131-a4ea-d4511a2364b7/26c1ff5b-9eda-4131-a4ea-d4511a2364b7.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 873.477602] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-99ca904a-8779-4fe2-8ac7-e964002d316e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.487917] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 873.487917] env[62692]: value = "task-1141238" [ 873.487917] env[62692]: _type = "Task" [ 873.487917] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.499075] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141238, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.640566] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.074s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.644103] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.262s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.646053] env[62692]: INFO nova.compute.claims [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 873.655594] env[62692]: DEBUG oslo_vmware.api [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141237, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.681126] env[62692]: INFO nova.scheduler.client.report [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Deleted allocations for instance ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba [ 873.804656] env[62692]: DEBUG nova.compute.manager [req-6e33c83a-d084-4bc1-a889-a7c568cd4fea req-fa2781ea-6e78-41bf-a4ce-4ae87676967d service nova] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Received event network-changed-318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 873.806147] env[62692]: DEBUG nova.compute.manager [req-6e33c83a-d084-4bc1-a889-a7c568cd4fea req-fa2781ea-6e78-41bf-a4ce-4ae87676967d service nova] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Refreshing instance network info cache due to event network-changed-318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 873.806263] env[62692]: DEBUG oslo_concurrency.lockutils [req-6e33c83a-d084-4bc1-a889-a7c568cd4fea req-fa2781ea-6e78-41bf-a4ce-4ae87676967d service nova] Acquiring lock "refresh_cache-6b2d3385-a9af-432b-a1e7-c91adb2e1f69" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.806467] env[62692]: DEBUG oslo_concurrency.lockutils [req-6e33c83a-d084-4bc1-a889-a7c568cd4fea req-fa2781ea-6e78-41bf-a4ce-4ae87676967d service nova] Acquired lock "refresh_cache-6b2d3385-a9af-432b-a1e7-c91adb2e1f69" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.806853] env[62692]: DEBUG nova.network.neutron [req-6e33c83a-d084-4bc1-a889-a7c568cd4fea req-fa2781ea-6e78-41bf-a4ce-4ae87676967d service nova] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Refreshing network info cache for port 318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 873.885279] env[62692]: DEBUG oslo_concurrency.lockutils [req-254e5b7d-4020-42b1-802f-330024c8db96 req-1663b8b3-6905-44eb-9bb3-c3f70874a325 service nova] Releasing lock "refresh_cache-754b3c65-1e4b-49d2-8980-095d975edb01" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.920175] env[62692]: DEBUG nova.compute.manager [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 873.966228] env[62692]: DEBUG oslo_vmware.api [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141232, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.998534] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141238, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.153644] env[62692]: DEBUG oslo_vmware.api [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141237, 'name': RemoveSnapshot_Task, 'duration_secs': 0.809092} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.153916] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Deleted Snapshot of the VM instance {{(pid=62692) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 874.154829] env[62692]: INFO nova.compute.manager [None req-553220a8-041b-4e02-8c41-9fdcdf610496 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Took 17.61 seconds to snapshot the instance on the hypervisor. [ 874.191275] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2550ffcc-35ec-4d24-b6c7-c45d802bb8f3 tempest-InstanceActionsNegativeTestJSON-1900842828 tempest-InstanceActionsNegativeTestJSON-1900842828-project-member] Lock "ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.936s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.396473] env[62692]: DEBUG nova.network.neutron [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Successfully updated port: 49431714-cc36-41a1-a843-809653e38ccc {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 874.444965] env[62692]: DEBUG oslo_concurrency.lockutils [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.455208] env[62692]: DEBUG nova.compute.manager [req-7e21df48-d957-43cd-b45e-4e96e75646ad req-16a82bf4-bca4-456c-a6d2-a99645b27842 service nova] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Received event network-vif-plugged-49431714-cc36-41a1-a843-809653e38ccc {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 874.455438] env[62692]: DEBUG oslo_concurrency.lockutils [req-7e21df48-d957-43cd-b45e-4e96e75646ad req-16a82bf4-bca4-456c-a6d2-a99645b27842 service nova] Acquiring lock "f44e2ee2-6321-4237-9644-a19b5e90ae16-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.455649] env[62692]: DEBUG oslo_concurrency.lockutils [req-7e21df48-d957-43cd-b45e-4e96e75646ad req-16a82bf4-bca4-456c-a6d2-a99645b27842 service nova] Lock "f44e2ee2-6321-4237-9644-a19b5e90ae16-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.455818] env[62692]: DEBUG oslo_concurrency.lockutils [req-7e21df48-d957-43cd-b45e-4e96e75646ad req-16a82bf4-bca4-456c-a6d2-a99645b27842 service nova] Lock "f44e2ee2-6321-4237-9644-a19b5e90ae16-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.455996] env[62692]: DEBUG nova.compute.manager [req-7e21df48-d957-43cd-b45e-4e96e75646ad req-16a82bf4-bca4-456c-a6d2-a99645b27842 service nova] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] No waiting events found dispatching network-vif-plugged-49431714-cc36-41a1-a843-809653e38ccc {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 874.456213] env[62692]: WARNING nova.compute.manager [req-7e21df48-d957-43cd-b45e-4e96e75646ad req-16a82bf4-bca4-456c-a6d2-a99645b27842 service nova] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Received unexpected event network-vif-plugged-49431714-cc36-41a1-a843-809653e38ccc for instance with vm_state building and task_state block_device_mapping. [ 874.456378] env[62692]: DEBUG nova.compute.manager [req-7e21df48-d957-43cd-b45e-4e96e75646ad req-16a82bf4-bca4-456c-a6d2-a99645b27842 service nova] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Received event network-changed-49431714-cc36-41a1-a843-809653e38ccc {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 874.456533] env[62692]: DEBUG nova.compute.manager [req-7e21df48-d957-43cd-b45e-4e96e75646ad req-16a82bf4-bca4-456c-a6d2-a99645b27842 service nova] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Refreshing instance network info cache due to event network-changed-49431714-cc36-41a1-a843-809653e38ccc. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 874.456716] env[62692]: DEBUG oslo_concurrency.lockutils [req-7e21df48-d957-43cd-b45e-4e96e75646ad req-16a82bf4-bca4-456c-a6d2-a99645b27842 service nova] Acquiring lock "refresh_cache-f44e2ee2-6321-4237-9644-a19b5e90ae16" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.456955] env[62692]: DEBUG oslo_concurrency.lockutils [req-7e21df48-d957-43cd-b45e-4e96e75646ad req-16a82bf4-bca4-456c-a6d2-a99645b27842 service nova] Acquired lock "refresh_cache-f44e2ee2-6321-4237-9644-a19b5e90ae16" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.457219] env[62692]: DEBUG nova.network.neutron [req-7e21df48-d957-43cd-b45e-4e96e75646ad req-16a82bf4-bca4-456c-a6d2-a99645b27842 service nova] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Refreshing network info cache for port 49431714-cc36-41a1-a843-809653e38ccc {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 874.470617] env[62692]: DEBUG oslo_vmware.api [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141232, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.499173] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141238, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.571462} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.499330] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 26c1ff5b-9eda-4131-a4ea-d4511a2364b7/26c1ff5b-9eda-4131-a4ea-d4511a2364b7.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 874.499517] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 874.499775] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0f8c8625-d87d-4077-b40c-7fd81b89622a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.507488] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 874.507488] env[62692]: value = "task-1141239" [ 874.507488] env[62692]: _type = "Task" [ 874.507488] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.521049] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141239, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.530886] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Acquiring lock "6b2d3385-a9af-432b-a1e7-c91adb2e1f69" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.531350] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Lock "6b2d3385-a9af-432b-a1e7-c91adb2e1f69" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.531661] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Acquiring lock "6b2d3385-a9af-432b-a1e7-c91adb2e1f69-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.531939] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Lock "6b2d3385-a9af-432b-a1e7-c91adb2e1f69-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.532190] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Lock "6b2d3385-a9af-432b-a1e7-c91adb2e1f69-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.535803] env[62692]: INFO nova.compute.manager [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Terminating instance [ 874.538382] env[62692]: DEBUG nova.compute.manager [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 874.538643] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 874.539803] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c55c80be-f54c-4740-a3ff-6ab57b41ec9f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.550043] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 874.551132] env[62692]: DEBUG nova.network.neutron [req-6e33c83a-d084-4bc1-a889-a7c568cd4fea req-fa2781ea-6e78-41bf-a4ce-4ae87676967d service nova] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Updated VIF entry in instance network info cache for port 318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 874.551573] env[62692]: DEBUG nova.network.neutron [req-6e33c83a-d084-4bc1-a889-a7c568cd4fea req-fa2781ea-6e78-41bf-a4ce-4ae87676967d service nova] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Updating instance_info_cache with network_info: [{"id": "318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db", "address": "fa:16:3e:32:9a:8d", "network": {"id": "7e7a1dcd-36ba-4a25-affa-6bd7fe895ade", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-508011613-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9667b0b7abaa4bf0a7e54e9d540df728", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap318dc9c9-0a", "ovs_interfaceid": "318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.552970] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-757fec3d-83c0-4026-96a5-7e65c5f410ce {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.561956] env[62692]: DEBUG oslo_vmware.api [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Waiting for the task: (returnval){ [ 874.561956] env[62692]: value = "task-1141240" [ 874.561956] env[62692]: _type = "Task" [ 874.561956] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.574508] env[62692]: DEBUG oslo_vmware.api [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1141240, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.902127] env[62692]: DEBUG oslo_concurrency.lockutils [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Acquiring lock "refresh_cache-f44e2ee2-6321-4237-9644-a19b5e90ae16" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.971883] env[62692]: DEBUG oslo_vmware.api [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141232, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.017979] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141239, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.104813} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.019213] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 875.019213] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1680e859-3752-4d1c-bd0c-11a1daed3606 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.024223] env[62692]: DEBUG nova.network.neutron [req-7e21df48-d957-43cd-b45e-4e96e75646ad req-16a82bf4-bca4-456c-a6d2-a99645b27842 service nova] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 875.048143] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] 26c1ff5b-9eda-4131-a4ea-d4511a2364b7/26c1ff5b-9eda-4131-a4ea-d4511a2364b7.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 875.049819] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-89196081-f678-49b3-af6a-2eea22a7696c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.066845] env[62692]: DEBUG oslo_concurrency.lockutils [req-6e33c83a-d084-4bc1-a889-a7c568cd4fea req-fa2781ea-6e78-41bf-a4ce-4ae87676967d service nova] Releasing lock "refresh_cache-6b2d3385-a9af-432b-a1e7-c91adb2e1f69" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.077699] env[62692]: DEBUG oslo_vmware.api [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1141240, 'name': PowerOffVM_Task, 'duration_secs': 0.217137} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.081289] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 875.081485] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 875.082327] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 875.082327] env[62692]: value = "task-1141241" [ 875.082327] env[62692]: _type = "Task" [ 875.082327] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.082717] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a3323e82-93f8-4ed6-8c9c-ae2318a503c6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.092862] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141241, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.115163] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20874cef-b9bb-4356-ac15-d8446fda7e53 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.125234] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3e4ce0c-9893-4ef7-8626-a68c3dea3fcf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.159388] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fa91713-002a-4f2e-8626-cdb11d2f8c7f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.163073] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 875.163073] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 875.163073] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Deleting the datastore file [datastore1] 6b2d3385-a9af-432b-a1e7-c91adb2e1f69 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 875.163073] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3a64cede-3b8d-475f-bd45-e19922e91d53 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.169982] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3789dc9-52eb-4ef7-81fe-c2af26eb798b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.177124] env[62692]: DEBUG oslo_vmware.api [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Waiting for the task: (returnval){ [ 875.177124] env[62692]: value = "task-1141243" [ 875.177124] env[62692]: _type = "Task" [ 875.177124] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.187083] env[62692]: DEBUG nova.compute.provider_tree [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 875.194887] env[62692]: DEBUG oslo_vmware.api [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1141243, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.200736] env[62692]: DEBUG nova.network.neutron [req-7e21df48-d957-43cd-b45e-4e96e75646ad req-16a82bf4-bca4-456c-a6d2-a99645b27842 service nova] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.207140] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquiring lock "fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.207396] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Lock "fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.397497] env[62692]: DEBUG nova.compute.manager [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 875.398074] env[62692]: DEBUG nova.virt.hardware [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 875.398309] env[62692]: DEBUG nova.virt.hardware [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 875.398468] env[62692]: DEBUG nova.virt.hardware [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 875.398657] env[62692]: DEBUG nova.virt.hardware [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 875.398835] env[62692]: DEBUG nova.virt.hardware [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 875.399020] env[62692]: DEBUG nova.virt.hardware [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 875.399247] env[62692]: DEBUG nova.virt.hardware [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 875.399417] env[62692]: DEBUG nova.virt.hardware [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 875.399587] env[62692]: DEBUG nova.virt.hardware [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 875.399763] env[62692]: DEBUG nova.virt.hardware [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 875.399935] env[62692]: DEBUG nova.virt.hardware [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 875.400858] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdbe3c28-069a-4080-b073-ee3f248f3bf1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.409781] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8ed2943-959a-4b18-9700-41c33a7f1e18 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.443148] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquiring lock "fac426b6-747a-455e-85a1-44e71b432180" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.443148] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Lock "fac426b6-747a-455e-85a1-44e71b432180" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.469661] env[62692]: DEBUG oslo_vmware.api [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141232, 'name': ReconfigVM_Task, 'duration_secs': 6.832523} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.472018] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "27583cb4-0478-44d8-8970-478682271513" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.472018] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Reconfigured VM to detach interface {{(pid=62692) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 875.596926] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141241, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.685751] env[62692]: DEBUG oslo_vmware.api [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1141243, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.20206} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.686052] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 875.686254] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 875.686523] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 875.686725] env[62692]: INFO nova.compute.manager [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Took 1.15 seconds to destroy the instance on the hypervisor. [ 875.687133] env[62692]: DEBUG oslo.service.loopingcall [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.687352] env[62692]: DEBUG nova.compute.manager [-] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 875.687456] env[62692]: DEBUG nova.network.neutron [-] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 875.689686] env[62692]: DEBUG nova.scheduler.client.report [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 875.707875] env[62692]: DEBUG oslo_concurrency.lockutils [req-7e21df48-d957-43cd-b45e-4e96e75646ad req-16a82bf4-bca4-456c-a6d2-a99645b27842 service nova] Releasing lock "refresh_cache-f44e2ee2-6321-4237-9644-a19b5e90ae16" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.707875] env[62692]: DEBUG oslo_concurrency.lockutils [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Acquired lock "refresh_cache-f44e2ee2-6321-4237-9644-a19b5e90ae16" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.707875] env[62692]: DEBUG nova.network.neutron [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 875.709262] env[62692]: DEBUG nova.compute.manager [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 876.101895] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141241, 'name': ReconfigVM_Task, 'duration_secs': 0.581311} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.102538] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Reconfigured VM instance instance-0000003c to attach disk [datastore1] 26c1ff5b-9eda-4131-a4ea-d4511a2364b7/26c1ff5b-9eda-4131-a4ea-d4511a2364b7.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 876.103512] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dde29fbc-887a-4a9a-9c87-b0e926d8f72a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.110839] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 876.110839] env[62692]: value = "task-1141244" [ 876.110839] env[62692]: _type = "Task" [ 876.110839] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.120385] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141244, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.195063] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.551s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.195682] env[62692]: DEBUG nova.compute.manager [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 876.198528] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 29.281s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.198762] env[62692]: DEBUG nova.objects.instance [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62692) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 876.225364] env[62692]: DEBUG nova.compute.manager [req-334a708a-074a-4f72-8477-14ca8844cefa req-19630be1-02ae-4a61-a7d3-f1c0b0ab96c4 service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] Received event network-vif-deleted-1b690180-b6c8-4a42-96de-cfd6f25d1791 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 876.225603] env[62692]: INFO nova.compute.manager [req-334a708a-074a-4f72-8477-14ca8844cefa req-19630be1-02ae-4a61-a7d3-f1c0b0ab96c4 service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] Neutron deleted interface 1b690180-b6c8-4a42-96de-cfd6f25d1791; detaching it from the instance and deleting it from the info cache [ 876.226023] env[62692]: DEBUG nova.network.neutron [req-334a708a-074a-4f72-8477-14ca8844cefa req-19630be1-02ae-4a61-a7d3-f1c0b0ab96c4 service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] Updating instance_info_cache with network_info: [{"id": "d4991635-5c62-4473-9661-97ac75a187fb", "address": "fa:16:3e:d4:98:43", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4991635-5c", "ovs_interfaceid": "d4991635-5c62-4473-9661-97ac75a187fb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.250021] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.290327] env[62692]: DEBUG nova.network.neutron [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 876.452437] env[62692]: DEBUG nova.network.neutron [-] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.591955] env[62692]: DEBUG nova.network.neutron [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Updating instance_info_cache with network_info: [{"id": "49431714-cc36-41a1-a843-809653e38ccc", "address": "fa:16:3e:cc:72:0e", "network": {"id": "c11d1069-c059-4b11-8d4e-f98ee75bab1e", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-928652414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7abef2aee3c4669801439f8e0fdad22", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "163e60bd-32d6-41c5-95e6-2eb10c5c9245", "external-id": "nsx-vlan-transportzone-716", "segmentation_id": 716, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap49431714-cc", "ovs_interfaceid": "49431714-cc36-41a1-a843-809653e38ccc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.606433] env[62692]: DEBUG nova.compute.manager [req-a3e4decc-958a-4371-a552-5ed8efccc913 req-4fa6171f-3648-440a-b718-52a441530922 service nova] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Received event network-vif-deleted-318dc9c9-0a9f-4f6b-af60-f9fe2ebe98db {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 876.622187] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141244, 'name': Rename_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.714216] env[62692]: DEBUG nova.compute.utils [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 876.716264] env[62692]: DEBUG nova.compute.manager [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 876.716585] env[62692]: DEBUG nova.network.neutron [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 876.735103] env[62692]: DEBUG oslo_concurrency.lockutils [req-334a708a-074a-4f72-8477-14ca8844cefa req-19630be1-02ae-4a61-a7d3-f1c0b0ab96c4 service nova] Acquiring lock "27583cb4-0478-44d8-8970-478682271513" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.735103] env[62692]: DEBUG oslo_concurrency.lockutils [req-334a708a-074a-4f72-8477-14ca8844cefa req-19630be1-02ae-4a61-a7d3-f1c0b0ab96c4 service nova] Acquired lock "27583cb4-0478-44d8-8970-478682271513" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.735103] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41d45c17-f11d-437b-b471-430e06ed2c8e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.755468] env[62692]: DEBUG oslo_concurrency.lockutils [req-334a708a-074a-4f72-8477-14ca8844cefa req-19630be1-02ae-4a61-a7d3-f1c0b0ab96c4 service nova] Releasing lock "27583cb4-0478-44d8-8970-478682271513" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.755856] env[62692]: WARNING nova.compute.manager [req-334a708a-074a-4f72-8477-14ca8844cefa req-19630be1-02ae-4a61-a7d3-f1c0b0ab96c4 service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] Detach interface failed, port_id=1b690180-b6c8-4a42-96de-cfd6f25d1791, reason: No device with interface-id 1b690180-b6c8-4a42-96de-cfd6f25d1791 exists on VM: nova.exception.NotFound: No device with interface-id 1b690180-b6c8-4a42-96de-cfd6f25d1791 exists on VM [ 876.774457] env[62692]: DEBUG nova.policy [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b40c8976b544c059458aa4e0dbab160', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9cb8ae08824245a7911de7455e7fcb68', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 876.876655] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "27583cb4-0478-44d8-8970-478682271513" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.876655] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "27583cb4-0478-44d8-8970-478682271513" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.876744] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "27583cb4-0478-44d8-8970-478682271513-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.876947] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "27583cb4-0478-44d8-8970-478682271513-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.877132] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "27583cb4-0478-44d8-8970-478682271513-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.879700] env[62692]: INFO nova.compute.manager [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Terminating instance [ 876.885391] env[62692]: DEBUG nova.compute.manager [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 876.885391] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 876.885391] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e0ddf09-dbb3-465a-ae46-3e9d17259cb7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.893858] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 876.894166] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-46de4e25-5a09-4f02-a572-f959dc609438 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.902472] env[62692]: DEBUG oslo_vmware.api [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 876.902472] env[62692]: value = "task-1141245" [ 876.902472] env[62692]: _type = "Task" [ 876.902472] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.911505] env[62692]: DEBUG oslo_vmware.api [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141245, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.957336] env[62692]: INFO nova.compute.manager [-] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Took 1.27 seconds to deallocate network for instance. [ 877.055123] env[62692]: DEBUG nova.network.neutron [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Successfully created port: 5cd6131b-792a-47dc-afcc-399d6be018ae {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 877.095354] env[62692]: DEBUG oslo_concurrency.lockutils [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Releasing lock "refresh_cache-f44e2ee2-6321-4237-9644-a19b5e90ae16" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.095698] env[62692]: DEBUG nova.compute.manager [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Instance network_info: |[{"id": "49431714-cc36-41a1-a843-809653e38ccc", "address": "fa:16:3e:cc:72:0e", "network": {"id": "c11d1069-c059-4b11-8d4e-f98ee75bab1e", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-928652414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7abef2aee3c4669801439f8e0fdad22", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "163e60bd-32d6-41c5-95e6-2eb10c5c9245", "external-id": "nsx-vlan-transportzone-716", "segmentation_id": 716, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap49431714-cc", "ovs_interfaceid": "49431714-cc36-41a1-a843-809653e38ccc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 877.096454] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cc:72:0e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '163e60bd-32d6-41c5-95e6-2eb10c5c9245', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '49431714-cc36-41a1-a843-809653e38ccc', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 877.104570] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Creating folder: Project (f7abef2aee3c4669801439f8e0fdad22). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 877.104826] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8940541f-4120-4e80-a510-faa2b9ae9ad0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.121717] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141244, 'name': Rename_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.123084] env[62692]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 877.123255] env[62692]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62692) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 877.123639] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Folder already exists: Project (f7abef2aee3c4669801439f8e0fdad22). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 877.123903] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Creating folder: Instances. Parent ref: group-v248891. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 877.124537] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1c2793fd-98ab-4a0f-bf5e-161e6b6c5eb0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.126976] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "refresh_cache-27583cb4-0478-44d8-8970-478682271513" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.127135] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "refresh_cache-27583cb4-0478-44d8-8970-478682271513" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.127315] env[62692]: DEBUG nova.network.neutron [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 877.139370] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Created folder: Instances in parent group-v248891. [ 877.140268] env[62692]: DEBUG oslo.service.loopingcall [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 877.140268] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 877.140268] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e7a8b84a-4fcc-4db5-b935-a6ada7dbe16a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.162925] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 877.162925] env[62692]: value = "task-1141248" [ 877.162925] env[62692]: _type = "Task" [ 877.162925] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.174018] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141248, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.217330] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8203a8fe-c96c-4d6b-8654-4f49e66b49dd tempest-ServersAdmin275Test-1648037221 tempest-ServersAdmin275Test-1648037221-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.019s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.220514] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.179s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.220781] env[62692]: DEBUG nova.objects.instance [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lazy-loading 'resources' on Instance uuid 8b75e781-930e-4885-81d2-8a8929d6c39b {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 877.227204] env[62692]: DEBUG nova.compute.manager [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 877.414075] env[62692]: DEBUG oslo_vmware.api [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141245, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.467391] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.624931] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141244, 'name': Rename_Task, 'duration_secs': 1.411909} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.624931] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 877.624931] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-babfc31b-6c6e-41ad-9029-6712e5883bbc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.635354] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 877.635354] env[62692]: value = "task-1141249" [ 877.635354] env[62692]: _type = "Task" [ 877.635354] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.646949] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141249, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.677545] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141248, 'name': CreateVM_Task, 'duration_secs': 0.390901} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.679916] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 877.680674] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'mount_device': '/dev/sda', 'attachment_id': 'd157af8f-5cb0-47ae-9b47-5742bdf3c1a5', 'guest_format': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-248898', 'volume_id': '4e85e9f0-749c-4394-8d51-fe35fa42cd6d', 'name': 'volume-4e85e9f0-749c-4394-8d51-fe35fa42cd6d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f44e2ee2-6321-4237-9644-a19b5e90ae16', 'attached_at': '', 'detached_at': '', 'volume_id': '4e85e9f0-749c-4394-8d51-fe35fa42cd6d', 'serial': '4e85e9f0-749c-4394-8d51-fe35fa42cd6d'}, 'device_type': None, 'delete_on_termination': True, 'disk_bus': None, 'boot_index': 0, 'volume_type': None}], 'swap': None} {{(pid=62692) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 877.680933] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Root volume attach. Driver type: vmdk {{(pid=62692) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 877.682088] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4f0fd3f-5a49-4d20-a686-b3642765ffd7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.691436] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f82db06-09ed-430e-8754-016d2952c391 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.700197] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0cb200a-54b0-4017-9429-7d48ee033915 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.707790] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-1a8ba1d4-cfc3-467b-bd34-58eee726e534 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.718410] env[62692]: DEBUG oslo_vmware.api [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Waiting for the task: (returnval){ [ 877.718410] env[62692]: value = "task-1141250" [ 877.718410] env[62692]: _type = "Task" [ 877.718410] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.726191] env[62692]: DEBUG oslo_vmware.api [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Task: {'id': task-1141250, 'name': RelocateVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.914780] env[62692]: DEBUG oslo_vmware.api [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141245, 'name': PowerOffVM_Task, 'duration_secs': 0.803101} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.915133] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 877.915260] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 877.915523] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cb97b854-44bc-4141-8de7-032cf5c9c757 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.992775] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 877.993218] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 877.993546] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Deleting the datastore file [datastore1] 27583cb4-0478-44d8-8970-478682271513 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 877.993935] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8a4039c6-a55e-4a7a-9373-298a603dfe68 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.006413] env[62692]: DEBUG oslo_vmware.api [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 878.006413] env[62692]: value = "task-1141252" [ 878.006413] env[62692]: _type = "Task" [ 878.006413] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.020293] env[62692]: DEBUG oslo_vmware.api [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141252, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.147952] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141249, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.217289] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12ce4a4c-f4c6-4dad-97b8-2087614800bb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.232519] env[62692]: DEBUG oslo_vmware.api [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Task: {'id': task-1141250, 'name': RelocateVM_Task} progress is 20%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.233614] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-844425c5-6837-4d24-aeb2-a94953160b9b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.238146] env[62692]: DEBUG nova.compute.manager [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 878.272197] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cfdd01e-1e31-45ec-b1f7-99aa4ca39405 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.282047] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24ca6b73-0694-429e-98cd-5726f4a549d0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.289558] env[62692]: DEBUG nova.network.neutron [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Updating instance_info_cache with network_info: [{"id": "d4991635-5c62-4473-9661-97ac75a187fb", "address": "fa:16:3e:d4:98:43", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4991635-5c", "ovs_interfaceid": "d4991635-5c62-4473-9661-97ac75a187fb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.304508] env[62692]: DEBUG nova.compute.provider_tree [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.321705] env[62692]: DEBUG nova.virt.hardware [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 878.322342] env[62692]: DEBUG nova.virt.hardware [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 878.322342] env[62692]: DEBUG nova.virt.hardware [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 878.323033] env[62692]: DEBUG nova.virt.hardware [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 878.323033] env[62692]: DEBUG nova.virt.hardware [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 878.323033] env[62692]: DEBUG nova.virt.hardware [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 878.323179] env[62692]: DEBUG nova.virt.hardware [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 878.323226] env[62692]: DEBUG nova.virt.hardware [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 878.323398] env[62692]: DEBUG nova.virt.hardware [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 878.323567] env[62692]: DEBUG nova.virt.hardware [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 878.323748] env[62692]: DEBUG nova.virt.hardware [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 878.324823] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e35e1b60-f077-41b2-a1ca-d06bd364a1d8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.334664] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc0093f4-f164-4204-a66f-fac0b13b4381 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.518507] env[62692]: DEBUG oslo_vmware.api [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141252, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.180788} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.518801] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 878.519031] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 878.519240] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 878.519425] env[62692]: INFO nova.compute.manager [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 27583cb4-0478-44d8-8970-478682271513] Took 1.64 seconds to destroy the instance on the hypervisor. [ 878.519809] env[62692]: DEBUG oslo.service.loopingcall [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.519946] env[62692]: DEBUG nova.compute.manager [-] [instance: 27583cb4-0478-44d8-8970-478682271513] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 878.519985] env[62692]: DEBUG nova.network.neutron [-] [instance: 27583cb4-0478-44d8-8970-478682271513] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 878.647842] env[62692]: DEBUG oslo_vmware.api [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141249, 'name': PowerOnVM_Task, 'duration_secs': 0.652417} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.648146] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 878.648361] env[62692]: INFO nova.compute.manager [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Took 15.35 seconds to spawn the instance on the hypervisor. [ 878.648554] env[62692]: DEBUG nova.compute.manager [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 878.649417] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44cc7625-e533-4d14-8776-05f8a53f7e77 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.666457] env[62692]: DEBUG nova.compute.manager [req-a965248a-ff29-48e4-a30b-e73788e20c9a req-a496596f-d3f7-418e-96b6-abef42391bee service nova] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Received event network-vif-plugged-5cd6131b-792a-47dc-afcc-399d6be018ae {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 878.666727] env[62692]: DEBUG oslo_concurrency.lockutils [req-a965248a-ff29-48e4-a30b-e73788e20c9a req-a496596f-d3f7-418e-96b6-abef42391bee service nova] Acquiring lock "433f740c-bde4-4d33-baca-5a7ac8179da1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.667121] env[62692]: DEBUG oslo_concurrency.lockutils [req-a965248a-ff29-48e4-a30b-e73788e20c9a req-a496596f-d3f7-418e-96b6-abef42391bee service nova] Lock "433f740c-bde4-4d33-baca-5a7ac8179da1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.667121] env[62692]: DEBUG oslo_concurrency.lockutils [req-a965248a-ff29-48e4-a30b-e73788e20c9a req-a496596f-d3f7-418e-96b6-abef42391bee service nova] Lock "433f740c-bde4-4d33-baca-5a7ac8179da1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.667228] env[62692]: DEBUG nova.compute.manager [req-a965248a-ff29-48e4-a30b-e73788e20c9a req-a496596f-d3f7-418e-96b6-abef42391bee service nova] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] No waiting events found dispatching network-vif-plugged-5cd6131b-792a-47dc-afcc-399d6be018ae {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 878.667399] env[62692]: WARNING nova.compute.manager [req-a965248a-ff29-48e4-a30b-e73788e20c9a req-a496596f-d3f7-418e-96b6-abef42391bee service nova] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Received unexpected event network-vif-plugged-5cd6131b-792a-47dc-afcc-399d6be018ae for instance with vm_state building and task_state spawning. [ 878.731067] env[62692]: DEBUG oslo_vmware.api [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Task: {'id': task-1141250, 'name': RelocateVM_Task, 'duration_secs': 0.522012} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.731415] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Volume attach. Driver type: vmdk {{(pid=62692) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 878.731631] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-248898', 'volume_id': '4e85e9f0-749c-4394-8d51-fe35fa42cd6d', 'name': 'volume-4e85e9f0-749c-4394-8d51-fe35fa42cd6d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f44e2ee2-6321-4237-9644-a19b5e90ae16', 'attached_at': '', 'detached_at': '', 'volume_id': '4e85e9f0-749c-4394-8d51-fe35fa42cd6d', 'serial': '4e85e9f0-749c-4394-8d51-fe35fa42cd6d'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 878.732682] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b01a4d1b-2f3d-4c23-9db1-05d7488679e3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.754108] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3e5091c-3bfe-40f4-bafc-51cf4d2309d8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.780017] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] volume-4e85e9f0-749c-4394-8d51-fe35fa42cd6d/volume-4e85e9f0-749c-4394-8d51-fe35fa42cd6d.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 878.780491] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3256a059-9fdd-4412-b9ce-aed444506e50 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.798679] env[62692]: DEBUG nova.network.neutron [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Successfully updated port: 5cd6131b-792a-47dc-afcc-399d6be018ae {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 878.805394] env[62692]: DEBUG oslo_vmware.api [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Waiting for the task: (returnval){ [ 878.805394] env[62692]: value = "task-1141253" [ 878.805394] env[62692]: _type = "Task" [ 878.805394] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.806521] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "refresh_cache-27583cb4-0478-44d8-8970-478682271513" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.813999] env[62692]: DEBUG nova.scheduler.client.report [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 878.825851] env[62692]: DEBUG oslo_vmware.api [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Task: {'id': task-1141253, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.825851] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3f79fb6d-13c9-4cf0-9192-d7dcace7dd66 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "interface-27583cb4-0478-44d8-8970-478682271513-1b690180-b6c8-4a42-96de-cfd6f25d1791" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 11.026s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.171883] env[62692]: INFO nova.compute.manager [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Took 42.99 seconds to build instance. [ 879.301375] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "refresh_cache-433f740c-bde4-4d33-baca-5a7ac8179da1" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.301516] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired lock "refresh_cache-433f740c-bde4-4d33-baca-5a7ac8179da1" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.301699] env[62692]: DEBUG nova.network.neutron [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 879.313282] env[62692]: DEBUG oslo_vmware.api [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Task: {'id': task-1141253, 'name': ReconfigVM_Task, 'duration_secs': 0.282538} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.317518] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Reconfigured VM instance instance-0000003d to attach disk [datastore2] volume-4e85e9f0-749c-4394-8d51-fe35fa42cd6d/volume-4e85e9f0-749c-4394-8d51-fe35fa42cd6d.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 879.324190] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.103s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.326105] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c5c9314c-4f11-4bca-9a6a-8c2a09f0e059 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.338944] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.393s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.340687] env[62692]: INFO nova.compute.claims [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 879.353856] env[62692]: DEBUG oslo_vmware.api [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Waiting for the task: (returnval){ [ 879.353856] env[62692]: value = "task-1141254" [ 879.353856] env[62692]: _type = "Task" [ 879.353856] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.364016] env[62692]: DEBUG oslo_vmware.api [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Task: {'id': task-1141254, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.380202] env[62692]: INFO nova.scheduler.client.report [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Deleted allocations for instance 8b75e781-930e-4885-81d2-8a8929d6c39b [ 879.484974] env[62692]: DEBUG nova.compute.manager [req-938eccfc-4b21-4aca-a134-e8dc1fbe1552 req-490fa7a3-68ca-489a-a396-5466309af375 service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] Received event network-vif-deleted-d4991635-5c62-4473-9661-97ac75a187fb {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 879.484974] env[62692]: INFO nova.compute.manager [req-938eccfc-4b21-4aca-a134-e8dc1fbe1552 req-490fa7a3-68ca-489a-a396-5466309af375 service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] Neutron deleted interface d4991635-5c62-4473-9661-97ac75a187fb; detaching it from the instance and deleting it from the info cache [ 879.485284] env[62692]: DEBUG nova.network.neutron [req-938eccfc-4b21-4aca-a134-e8dc1fbe1552 req-490fa7a3-68ca-489a-a396-5466309af375 service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.673662] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a58926d1-818d-4e8d-a31e-5991c3be8bdf tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "26c1ff5b-9eda-4131-a4ea-d4511a2364b7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.846s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.867743] env[62692]: DEBUG oslo_vmware.api [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Task: {'id': task-1141254, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.870268] env[62692]: DEBUG nova.network.neutron [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 879.893460] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8b638d79-b642-44d7-90af-d88e53c25c6f tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "8b75e781-930e-4885-81d2-8a8929d6c39b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.305s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.917118] env[62692]: DEBUG nova.network.neutron [-] [instance: 27583cb4-0478-44d8-8970-478682271513] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.989708] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-570a0e58-77f5-489d-a689-8f02509c2a1d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.007418] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f64e5a04-1594-42d7-90b4-7d9d01556c55 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.050119] env[62692]: DEBUG nova.compute.manager [req-938eccfc-4b21-4aca-a134-e8dc1fbe1552 req-490fa7a3-68ca-489a-a396-5466309af375 service nova] [instance: 27583cb4-0478-44d8-8970-478682271513] Detach interface failed, port_id=d4991635-5c62-4473-9661-97ac75a187fb, reason: Instance 27583cb4-0478-44d8-8970-478682271513 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 880.092478] env[62692]: DEBUG nova.network.neutron [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Updating instance_info_cache with network_info: [{"id": "5cd6131b-792a-47dc-afcc-399d6be018ae", "address": "fa:16:3e:55:28:27", "network": {"id": "f6352999-f943-431d-b54a-c9dc1d810a85", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-121037111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9cb8ae08824245a7911de7455e7fcb68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c842425c-544e-4ce2-9657-512723bd318e", "external-id": "nsx-vlan-transportzone-80", "segmentation_id": 80, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5cd6131b-79", "ovs_interfaceid": "5cd6131b-792a-47dc-afcc-399d6be018ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.179588] env[62692]: DEBUG nova.compute.manager [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 880.367950] env[62692]: DEBUG oslo_vmware.api [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Task: {'id': task-1141254, 'name': ReconfigVM_Task, 'duration_secs': 0.796676} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.368598] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-248898', 'volume_id': '4e85e9f0-749c-4394-8d51-fe35fa42cd6d', 'name': 'volume-4e85e9f0-749c-4394-8d51-fe35fa42cd6d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f44e2ee2-6321-4237-9644-a19b5e90ae16', 'attached_at': '', 'detached_at': '', 'volume_id': '4e85e9f0-749c-4394-8d51-fe35fa42cd6d', 'serial': '4e85e9f0-749c-4394-8d51-fe35fa42cd6d'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 880.369909] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c1466b29-7d3f-40d4-9e74-ddc338cc559b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.380175] env[62692]: DEBUG oslo_vmware.api [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Waiting for the task: (returnval){ [ 880.380175] env[62692]: value = "task-1141255" [ 880.380175] env[62692]: _type = "Task" [ 880.380175] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.392656] env[62692]: DEBUG oslo_vmware.api [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Task: {'id': task-1141255, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.421989] env[62692]: INFO nova.compute.manager [-] [instance: 27583cb4-0478-44d8-8970-478682271513] Took 1.90 seconds to deallocate network for instance. [ 880.596175] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Releasing lock "refresh_cache-433f740c-bde4-4d33-baca-5a7ac8179da1" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.596478] env[62692]: DEBUG nova.compute.manager [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Instance network_info: |[{"id": "5cd6131b-792a-47dc-afcc-399d6be018ae", "address": "fa:16:3e:55:28:27", "network": {"id": "f6352999-f943-431d-b54a-c9dc1d810a85", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-121037111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9cb8ae08824245a7911de7455e7fcb68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c842425c-544e-4ce2-9657-512723bd318e", "external-id": "nsx-vlan-transportzone-80", "segmentation_id": 80, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5cd6131b-79", "ovs_interfaceid": "5cd6131b-792a-47dc-afcc-399d6be018ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 880.597058] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:55:28:27', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c842425c-544e-4ce2-9657-512723bd318e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5cd6131b-792a-47dc-afcc-399d6be018ae', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 880.607607] env[62692]: DEBUG oslo.service.loopingcall [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 880.612923] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 880.613942] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a2ff178e-beef-4424-8ea8-79d6359519a4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.644667] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 880.644667] env[62692]: value = "task-1141256" [ 880.644667] env[62692]: _type = "Task" [ 880.644667] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.659938] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141256, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.703424] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.845888] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67dfdc55-d578-41c0-a9a6-d247b1da38bc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.857745] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dba5d60a-56dc-4399-99be-45895c33bf87 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.907333] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aa11ec9-ee20-4d7c-8389-dab845b81823 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.916904] env[62692]: DEBUG oslo_vmware.api [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Task: {'id': task-1141255, 'name': Rename_Task, 'duration_secs': 0.152482} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.918646] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 880.918963] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a017a268-849a-46dd-aac8-1b74804426ca {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.922511] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4928bab0-8859-4276-9f57-e925dc9e49b1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.930287] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.938595] env[62692]: DEBUG nova.compute.provider_tree [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 880.944410] env[62692]: DEBUG oslo_vmware.api [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Waiting for the task: (returnval){ [ 880.944410] env[62692]: value = "task-1141257" [ 880.944410] env[62692]: _type = "Task" [ 880.944410] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.955468] env[62692]: DEBUG oslo_vmware.api [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Task: {'id': task-1141257, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.971893] env[62692]: DEBUG oslo_concurrency.lockutils [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.972447] env[62692]: DEBUG oslo_concurrency.lockutils [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.032460] env[62692]: DEBUG nova.compute.manager [req-d9021bfe-e570-4c02-a070-1c71b9279932 req-0f533ef7-529c-4b98-be5c-f356dce951fb service nova] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Received event network-changed-5cd6131b-792a-47dc-afcc-399d6be018ae {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 881.032670] env[62692]: DEBUG nova.compute.manager [req-d9021bfe-e570-4c02-a070-1c71b9279932 req-0f533ef7-529c-4b98-be5c-f356dce951fb service nova] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Refreshing instance network info cache due to event network-changed-5cd6131b-792a-47dc-afcc-399d6be018ae. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 881.032896] env[62692]: DEBUG oslo_concurrency.lockutils [req-d9021bfe-e570-4c02-a070-1c71b9279932 req-0f533ef7-529c-4b98-be5c-f356dce951fb service nova] Acquiring lock "refresh_cache-433f740c-bde4-4d33-baca-5a7ac8179da1" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.033666] env[62692]: DEBUG oslo_concurrency.lockutils [req-d9021bfe-e570-4c02-a070-1c71b9279932 req-0f533ef7-529c-4b98-be5c-f356dce951fb service nova] Acquired lock "refresh_cache-433f740c-bde4-4d33-baca-5a7ac8179da1" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.033793] env[62692]: DEBUG nova.network.neutron [req-d9021bfe-e570-4c02-a070-1c71b9279932 req-0f533ef7-529c-4b98-be5c-f356dce951fb service nova] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Refreshing network info cache for port 5cd6131b-792a-47dc-afcc-399d6be018ae {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 881.159122] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141256, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.373342] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "b23d3484-4fc5-46b2-8e8c-bdaab087f046" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.373342] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "b23d3484-4fc5-46b2-8e8c-bdaab087f046" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.446293] env[62692]: DEBUG nova.scheduler.client.report [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 881.465043] env[62692]: DEBUG oslo_vmware.api [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Task: {'id': task-1141257, 'name': PowerOnVM_Task, 'duration_secs': 0.519424} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.465201] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 881.465375] env[62692]: INFO nova.compute.manager [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Took 6.07 seconds to spawn the instance on the hypervisor. [ 881.465555] env[62692]: DEBUG nova.compute.manager [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 881.468613] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76eb29ac-ee03-4d79-b279-5df0dc3e1f42 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.666862] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141256, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.853888] env[62692]: DEBUG nova.network.neutron [req-d9021bfe-e570-4c02-a070-1c71b9279932 req-0f533ef7-529c-4b98-be5c-f356dce951fb service nova] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Updated VIF entry in instance network info cache for port 5cd6131b-792a-47dc-afcc-399d6be018ae. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 881.855158] env[62692]: DEBUG nova.network.neutron [req-d9021bfe-e570-4c02-a070-1c71b9279932 req-0f533ef7-529c-4b98-be5c-f356dce951fb service nova] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Updating instance_info_cache with network_info: [{"id": "5cd6131b-792a-47dc-afcc-399d6be018ae", "address": "fa:16:3e:55:28:27", "network": {"id": "f6352999-f943-431d-b54a-c9dc1d810a85", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-121037111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9cb8ae08824245a7911de7455e7fcb68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c842425c-544e-4ce2-9657-512723bd318e", "external-id": "nsx-vlan-transportzone-80", "segmentation_id": 80, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5cd6131b-79", "ovs_interfaceid": "5cd6131b-792a-47dc-afcc-399d6be018ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.958715] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.619s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.959185] env[62692]: DEBUG nova.compute.manager [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 881.967320] env[62692]: DEBUG oslo_concurrency.lockutils [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 30.253s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.995986] env[62692]: INFO nova.compute.manager [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Took 42.23 seconds to build instance. [ 882.158433] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141256, 'name': CreateVM_Task, 'duration_secs': 1.421136} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.158608] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 882.159375] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.159547] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.161260] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 882.161260] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0785ed5d-8954-4a20-a6ce-5c8b86cf93bb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.166557] env[62692]: DEBUG oslo_vmware.api [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 882.166557] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]521807e7-fd0a-3135-d7e0-f1699bbfac7b" [ 882.166557] env[62692]: _type = "Task" [ 882.166557] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.174650] env[62692]: DEBUG oslo_vmware.api [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]521807e7-fd0a-3135-d7e0-f1699bbfac7b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.272600] env[62692]: DEBUG oslo_concurrency.lockutils [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Acquiring lock "8a56e6df-eea1-41f4-9360-4f06d2f516a3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.273413] env[62692]: DEBUG oslo_concurrency.lockutils [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Lock "8a56e6df-eea1-41f4-9360-4f06d2f516a3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.361182] env[62692]: DEBUG oslo_concurrency.lockutils [req-d9021bfe-e570-4c02-a070-1c71b9279932 req-0f533ef7-529c-4b98-be5c-f356dce951fb service nova] Releasing lock "refresh_cache-433f740c-bde4-4d33-baca-5a7ac8179da1" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.470410] env[62692]: DEBUG nova.compute.utils [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 882.478309] env[62692]: INFO nova.compute.claims [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 882.482148] env[62692]: DEBUG nova.compute.manager [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 882.482148] env[62692]: DEBUG nova.network.neutron [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 882.500507] env[62692]: DEBUG oslo_concurrency.lockutils [None req-692342c6-5cc9-4b43-87fc-b2dec94e2880 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Lock "f44e2ee2-6321-4237-9644-a19b5e90ae16" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.204s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.605342] env[62692]: DEBUG nova.compute.manager [req-3ff9a00b-2cd9-4d31-a160-da19e1e48919 req-043e7630-30bc-473b-9c81-2e37fc25b3b7 service nova] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Received event network-changed-49431714-cc36-41a1-a843-809653e38ccc {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 882.606381] env[62692]: DEBUG nova.compute.manager [req-3ff9a00b-2cd9-4d31-a160-da19e1e48919 req-043e7630-30bc-473b-9c81-2e37fc25b3b7 service nova] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Refreshing instance network info cache due to event network-changed-49431714-cc36-41a1-a843-809653e38ccc. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 882.609448] env[62692]: DEBUG oslo_concurrency.lockutils [req-3ff9a00b-2cd9-4d31-a160-da19e1e48919 req-043e7630-30bc-473b-9c81-2e37fc25b3b7 service nova] Acquiring lock "refresh_cache-f44e2ee2-6321-4237-9644-a19b5e90ae16" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.612528] env[62692]: DEBUG oslo_concurrency.lockutils [req-3ff9a00b-2cd9-4d31-a160-da19e1e48919 req-043e7630-30bc-473b-9c81-2e37fc25b3b7 service nova] Acquired lock "refresh_cache-f44e2ee2-6321-4237-9644-a19b5e90ae16" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.612528] env[62692]: DEBUG nova.network.neutron [req-3ff9a00b-2cd9-4d31-a160-da19e1e48919 req-043e7630-30bc-473b-9c81-2e37fc25b3b7 service nova] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Refreshing network info cache for port 49431714-cc36-41a1-a843-809653e38ccc {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 882.682557] env[62692]: DEBUG oslo_vmware.api [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]521807e7-fd0a-3135-d7e0-f1699bbfac7b, 'name': SearchDatastore_Task, 'duration_secs': 0.011538} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.683131] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.683758] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 882.684169] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.684687] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.684936] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 882.685499] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-27ea3c1c-35a9-4c2a-b645-ef9c47e41b8a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.697857] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 882.698635] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 882.699965] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-022eb1df-36ce-4d68-82e2-0c75de73ff28 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.707881] env[62692]: DEBUG oslo_vmware.api [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 882.707881] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5214df80-5871-d6cf-3659-c7e1fb22df46" [ 882.707881] env[62692]: _type = "Task" [ 882.707881] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.718036] env[62692]: DEBUG oslo_vmware.api [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5214df80-5871-d6cf-3659-c7e1fb22df46, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.761589] env[62692]: DEBUG nova.policy [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b224b59c9c2b4870bcabc211a2ee7bc8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '951365ebda7a4c9e9455ab64e302ac8d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 882.987930] env[62692]: INFO nova.compute.resource_tracker [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Updating resource usage from migration 82adcbf0-fa69-4f63-b376-ff031ef9c49e [ 882.990458] env[62692]: DEBUG nova.compute.manager [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 883.005498] env[62692]: DEBUG nova.compute.manager [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 883.076166] env[62692]: DEBUG nova.network.neutron [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Successfully created port: 54f5eeb9-6d87-4308-94cd-8441cd77972b {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 883.221703] env[62692]: DEBUG oslo_vmware.api [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5214df80-5871-d6cf-3659-c7e1fb22df46, 'name': SearchDatastore_Task, 'duration_secs': 0.010351} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.222546] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ceface94-d760-4706-ad36-e0bad89bb328 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.238024] env[62692]: DEBUG oslo_vmware.api [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 883.238024] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52e520e2-99dd-6768-a139-9d878eb4087b" [ 883.238024] env[62692]: _type = "Task" [ 883.238024] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.254599] env[62692]: DEBUG oslo_vmware.api [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52e520e2-99dd-6768-a139-9d878eb4087b, 'name': SearchDatastore_Task, 'duration_secs': 0.011149} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.254735] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.254939] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 433f740c-bde4-4d33-baca-5a7ac8179da1/433f740c-bde4-4d33-baca-5a7ac8179da1.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 883.257516] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-be0510c0-9728-4528-ad5b-2db3f4222d6f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.266398] env[62692]: DEBUG oslo_vmware.api [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 883.266398] env[62692]: value = "task-1141258" [ 883.266398] env[62692]: _type = "Task" [ 883.266398] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.277532] env[62692]: DEBUG oslo_vmware.api [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141258, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.504021] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fae84411-5a80-44dc-a70e-4fbaed7f1f07 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.507797] env[62692]: DEBUG nova.network.neutron [req-3ff9a00b-2cd9-4d31-a160-da19e1e48919 req-043e7630-30bc-473b-9c81-2e37fc25b3b7 service nova] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Updated VIF entry in instance network info cache for port 49431714-cc36-41a1-a843-809653e38ccc. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 883.508648] env[62692]: DEBUG nova.network.neutron [req-3ff9a00b-2cd9-4d31-a160-da19e1e48919 req-043e7630-30bc-473b-9c81-2e37fc25b3b7 service nova] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Updating instance_info_cache with network_info: [{"id": "49431714-cc36-41a1-a843-809653e38ccc", "address": "fa:16:3e:cc:72:0e", "network": {"id": "c11d1069-c059-4b11-8d4e-f98ee75bab1e", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-928652414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.244", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7abef2aee3c4669801439f8e0fdad22", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "163e60bd-32d6-41c5-95e6-2eb10c5c9245", "external-id": "nsx-vlan-transportzone-716", "segmentation_id": 716, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap49431714-cc", "ovs_interfaceid": "49431714-cc36-41a1-a843-809653e38ccc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.518826] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6bea5fb-2536-404f-8e9b-7b9c806818f2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.556586] env[62692]: DEBUG oslo_concurrency.lockutils [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.557779] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a81b200-47d1-4d61-a5dd-9b03d8f79205 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.568437] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1593b64f-4623-4c52-8e35-0d4c0ce893fa {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.588655] env[62692]: DEBUG nova.compute.provider_tree [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 883.778316] env[62692]: DEBUG oslo_vmware.api [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141258, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.499848} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.778657] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 433f740c-bde4-4d33-baca-5a7ac8179da1/433f740c-bde4-4d33-baca-5a7ac8179da1.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 883.778883] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 883.779225] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-582566ad-f1c3-4228-bed8-d3750c21c96f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.786830] env[62692]: DEBUG oslo_vmware.api [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 883.786830] env[62692]: value = "task-1141259" [ 883.786830] env[62692]: _type = "Task" [ 883.786830] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.797035] env[62692]: DEBUG oslo_vmware.api [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141259, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.001319] env[62692]: DEBUG nova.compute.manager [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 884.013131] env[62692]: DEBUG oslo_concurrency.lockutils [req-3ff9a00b-2cd9-4d31-a160-da19e1e48919 req-043e7630-30bc-473b-9c81-2e37fc25b3b7 service nova] Releasing lock "refresh_cache-f44e2ee2-6321-4237-9644-a19b5e90ae16" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.028541] env[62692]: DEBUG nova.virt.hardware [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 884.028840] env[62692]: DEBUG nova.virt.hardware [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 884.029050] env[62692]: DEBUG nova.virt.hardware [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 884.029266] env[62692]: DEBUG nova.virt.hardware [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 884.029421] env[62692]: DEBUG nova.virt.hardware [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 884.029577] env[62692]: DEBUG nova.virt.hardware [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 884.029809] env[62692]: DEBUG nova.virt.hardware [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 884.030014] env[62692]: DEBUG nova.virt.hardware [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 884.030228] env[62692]: DEBUG nova.virt.hardware [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 884.030402] env[62692]: DEBUG nova.virt.hardware [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 884.030579] env[62692]: DEBUG nova.virt.hardware [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 884.031492] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e87d46ab-ea6c-4adb-a7ea-ba3686d5fa8d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.040075] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b630561a-61b5-4532-95fa-718fda504029 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.091853] env[62692]: DEBUG nova.scheduler.client.report [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 884.298025] env[62692]: DEBUG oslo_vmware.api [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141259, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069299} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.298324] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 884.299178] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a359251-1d37-4d61-bf5e-9dca53795f44 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.322461] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 433f740c-bde4-4d33-baca-5a7ac8179da1/433f740c-bde4-4d33-baca-5a7ac8179da1.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 884.322790] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1aed2fc0-c5b3-436c-91a3-df3376651374 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.344853] env[62692]: DEBUG oslo_vmware.api [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 884.344853] env[62692]: value = "task-1141260" [ 884.344853] env[62692]: _type = "Task" [ 884.344853] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.353509] env[62692]: DEBUG oslo_vmware.api [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141260, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.600021] env[62692]: DEBUG oslo_concurrency.lockutils [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.631s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.600021] env[62692]: INFO nova.compute.manager [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Migrating [ 884.600021] env[62692]: DEBUG oslo_concurrency.lockutils [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.600021] env[62692]: DEBUG oslo_concurrency.lockutils [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquired lock "compute-rpcapi-router" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.600021] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.434s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.600021] env[62692]: DEBUG nova.objects.instance [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Lazy-loading 'resources' on Instance uuid 7a3f9c7d-03cf-4177-88db-acf33cc03feb {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 884.731570] env[62692]: DEBUG nova.network.neutron [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Successfully updated port: 54f5eeb9-6d87-4308-94cd-8441cd77972b {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 884.814344] env[62692]: DEBUG nova.compute.manager [req-a6684811-fede-4da9-a3d9-707ba3c3d28b req-e345d5ae-bc06-4a77-85ad-3067f7037e86 service nova] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Received event network-vif-plugged-54f5eeb9-6d87-4308-94cd-8441cd77972b {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 884.814657] env[62692]: DEBUG oslo_concurrency.lockutils [req-a6684811-fede-4da9-a3d9-707ba3c3d28b req-e345d5ae-bc06-4a77-85ad-3067f7037e86 service nova] Acquiring lock "b389b14d-6dcc-4b26-84f8-38b952bf6ecd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.815302] env[62692]: DEBUG oslo_concurrency.lockutils [req-a6684811-fede-4da9-a3d9-707ba3c3d28b req-e345d5ae-bc06-4a77-85ad-3067f7037e86 service nova] Lock "b389b14d-6dcc-4b26-84f8-38b952bf6ecd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.815548] env[62692]: DEBUG oslo_concurrency.lockutils [req-a6684811-fede-4da9-a3d9-707ba3c3d28b req-e345d5ae-bc06-4a77-85ad-3067f7037e86 service nova] Lock "b389b14d-6dcc-4b26-84f8-38b952bf6ecd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.815812] env[62692]: DEBUG nova.compute.manager [req-a6684811-fede-4da9-a3d9-707ba3c3d28b req-e345d5ae-bc06-4a77-85ad-3067f7037e86 service nova] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] No waiting events found dispatching network-vif-plugged-54f5eeb9-6d87-4308-94cd-8441cd77972b {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 884.816076] env[62692]: WARNING nova.compute.manager [req-a6684811-fede-4da9-a3d9-707ba3c3d28b req-e345d5ae-bc06-4a77-85ad-3067f7037e86 service nova] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Received unexpected event network-vif-plugged-54f5eeb9-6d87-4308-94cd-8441cd77972b for instance with vm_state building and task_state spawning. [ 884.816301] env[62692]: DEBUG nova.compute.manager [req-a6684811-fede-4da9-a3d9-707ba3c3d28b req-e345d5ae-bc06-4a77-85ad-3067f7037e86 service nova] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Received event network-changed-54f5eeb9-6d87-4308-94cd-8441cd77972b {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 884.816476] env[62692]: DEBUG nova.compute.manager [req-a6684811-fede-4da9-a3d9-707ba3c3d28b req-e345d5ae-bc06-4a77-85ad-3067f7037e86 service nova] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Refreshing instance network info cache due to event network-changed-54f5eeb9-6d87-4308-94cd-8441cd77972b. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 884.816735] env[62692]: DEBUG oslo_concurrency.lockutils [req-a6684811-fede-4da9-a3d9-707ba3c3d28b req-e345d5ae-bc06-4a77-85ad-3067f7037e86 service nova] Acquiring lock "refresh_cache-b389b14d-6dcc-4b26-84f8-38b952bf6ecd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.816879] env[62692]: DEBUG oslo_concurrency.lockutils [req-a6684811-fede-4da9-a3d9-707ba3c3d28b req-e345d5ae-bc06-4a77-85ad-3067f7037e86 service nova] Acquired lock "refresh_cache-b389b14d-6dcc-4b26-84f8-38b952bf6ecd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.817133] env[62692]: DEBUG nova.network.neutron [req-a6684811-fede-4da9-a3d9-707ba3c3d28b req-e345d5ae-bc06-4a77-85ad-3067f7037e86 service nova] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Refreshing network info cache for port 54f5eeb9-6d87-4308-94cd-8441cd77972b {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 884.855374] env[62692]: DEBUG oslo_vmware.api [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141260, 'name': ReconfigVM_Task, 'duration_secs': 0.290272} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.855374] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 433f740c-bde4-4d33-baca-5a7ac8179da1/433f740c-bde4-4d33-baca-5a7ac8179da1.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 884.855802] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b217deda-ec52-4d35-82ae-dd518058f213 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.863443] env[62692]: DEBUG oslo_vmware.api [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 884.863443] env[62692]: value = "task-1141261" [ 884.863443] env[62692]: _type = "Task" [ 884.863443] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.874132] env[62692]: DEBUG oslo_vmware.api [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141261, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.102015] env[62692]: INFO nova.compute.rpcapi [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 885.102937] env[62692]: DEBUG oslo_concurrency.lockutils [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Releasing lock "compute-rpcapi-router" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.234458] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Acquiring lock "refresh_cache-b389b14d-6dcc-4b26-84f8-38b952bf6ecd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.349411] env[62692]: DEBUG nova.network.neutron [req-a6684811-fede-4da9-a3d9-707ba3c3d28b req-e345d5ae-bc06-4a77-85ad-3067f7037e86 service nova] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 885.375698] env[62692]: DEBUG oslo_vmware.api [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141261, 'name': Rename_Task, 'duration_secs': 0.134669} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.376541] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 885.376541] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-30a550d7-e617-47fe-839e-290f08f67a74 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.385308] env[62692]: DEBUG oslo_vmware.api [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 885.385308] env[62692]: value = "task-1141262" [ 885.385308] env[62692]: _type = "Task" [ 885.385308] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.399718] env[62692]: DEBUG oslo_vmware.api [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141262, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.462952] env[62692]: DEBUG nova.network.neutron [req-a6684811-fede-4da9-a3d9-707ba3c3d28b req-e345d5ae-bc06-4a77-85ad-3067f7037e86 service nova] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.481027] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4745e7e-5673-4dfc-9adf-405bbf08e1cc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.492045] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7e0b526-ac2e-463d-9d3e-7b3122cbbc7e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.529136] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ae87960-3ac9-4c7c-8c18-3469780cfdaa {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.538018] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5440ad63-7fbe-455d-9fa2-2971deced5e1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.553308] env[62692]: DEBUG nova.compute.provider_tree [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 885.620428] env[62692]: DEBUG oslo_concurrency.lockutils [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquiring lock "refresh_cache-2ccb76cf-f641-4306-a137-fb2417285df9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.620690] env[62692]: DEBUG oslo_concurrency.lockutils [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquired lock "refresh_cache-2ccb76cf-f641-4306-a137-fb2417285df9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.621088] env[62692]: DEBUG nova.network.neutron [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 885.897670] env[62692]: DEBUG oslo_vmware.api [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141262, 'name': PowerOnVM_Task} progress is 76%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.965915] env[62692]: DEBUG oslo_concurrency.lockutils [req-a6684811-fede-4da9-a3d9-707ba3c3d28b req-e345d5ae-bc06-4a77-85ad-3067f7037e86 service nova] Releasing lock "refresh_cache-b389b14d-6dcc-4b26-84f8-38b952bf6ecd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.966510] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Acquired lock "refresh_cache-b389b14d-6dcc-4b26-84f8-38b952bf6ecd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.966667] env[62692]: DEBUG nova.network.neutron [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 886.056562] env[62692]: DEBUG nova.scheduler.client.report [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 886.375181] env[62692]: DEBUG nova.network.neutron [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Updating instance_info_cache with network_info: [{"id": "951b0add-cfaa-4506-a6ac-56bcb8d1041e", "address": "fa:16:3e:70:90:85", "network": {"id": "a74c4d30-3568-4e50-98f5-d2fcc05ec728", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.89", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "989968e8796a4a70a4dbd045418a475f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap951b0add-cf", "ovs_interfaceid": "951b0add-cfaa-4506-a6ac-56bcb8d1041e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.396647] env[62692]: DEBUG oslo_vmware.api [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141262, 'name': PowerOnVM_Task, 'duration_secs': 0.990885} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.396933] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 886.397160] env[62692]: INFO nova.compute.manager [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Took 8.16 seconds to spawn the instance on the hypervisor. [ 886.397418] env[62692]: DEBUG nova.compute.manager [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 886.398227] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cce1539-2762-4e86-a1ce-69fe5b4410d0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.504602] env[62692]: DEBUG nova.network.neutron [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 886.562842] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.963s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.565352] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 32.968s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.585335] env[62692]: INFO nova.scheduler.client.report [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Deleted allocations for instance 7a3f9c7d-03cf-4177-88db-acf33cc03feb [ 886.679369] env[62692]: DEBUG nova.network.neutron [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Updating instance_info_cache with network_info: [{"id": "54f5eeb9-6d87-4308-94cd-8441cd77972b", "address": "fa:16:3e:27:cc:d2", "network": {"id": "e21c6ed1-5058-4943-b1a8-75627d2db38e", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1952046498-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "951365ebda7a4c9e9455ab64e302ac8d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap54f5eeb9-6d", "ovs_interfaceid": "54f5eeb9-6d87-4308-94cd-8441cd77972b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.878556] env[62692]: DEBUG oslo_concurrency.lockutils [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Releasing lock "refresh_cache-2ccb76cf-f641-4306-a137-fb2417285df9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.918533] env[62692]: INFO nova.compute.manager [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Took 41.55 seconds to build instance. [ 887.095642] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a13c0150-2a7e-493d-a5ec-87aa62bef48e tempest-ServersAdmin275Test-462179424 tempest-ServersAdmin275Test-462179424-project-member] Lock "7a3f9c7d-03cf-4177-88db-acf33cc03feb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.235s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.181765] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Releasing lock "refresh_cache-b389b14d-6dcc-4b26-84f8-38b952bf6ecd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.182348] env[62692]: DEBUG nova.compute.manager [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Instance network_info: |[{"id": "54f5eeb9-6d87-4308-94cd-8441cd77972b", "address": "fa:16:3e:27:cc:d2", "network": {"id": "e21c6ed1-5058-4943-b1a8-75627d2db38e", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1952046498-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "951365ebda7a4c9e9455ab64e302ac8d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap54f5eeb9-6d", "ovs_interfaceid": "54f5eeb9-6d87-4308-94cd-8441cd77972b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 887.182877] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:27:cc:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f3a2eb5-353f-45c5-a73b-869626f4bb13', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '54f5eeb9-6d87-4308-94cd-8441cd77972b', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 887.190901] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Creating folder: Project (951365ebda7a4c9e9455ab64e302ac8d). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 887.191456] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0b6e0e1d-6704-4a9e-beef-405f037b3f0e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.205432] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Created folder: Project (951365ebda7a4c9e9455ab64e302ac8d) in parent group-v248868. [ 887.205841] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Creating folder: Instances. Parent ref: group-v248966. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 887.206132] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f9216b71-9c79-4c1d-8e1b-c74cdb5847e9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.219089] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Created folder: Instances in parent group-v248966. [ 887.219461] env[62692]: DEBUG oslo.service.loopingcall [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 887.219686] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 887.219951] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-20f6ce51-d1f5-46a8-b918-14f995df5e14 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.244553] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 887.244553] env[62692]: value = "task-1141265" [ 887.244553] env[62692]: _type = "Task" [ 887.244553] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.257939] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141265, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.424991] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3909bb55-e6b7-4e85-8026-3c56fab6a3d0 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "433f740c-bde4-4d33-baca-5a7ac8179da1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.387s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.588437] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Applying migration context for instance 2ccb76cf-f641-4306-a137-fb2417285df9 as it has an incoming, in-progress migration 82adcbf0-fa69-4f63-b376-ff031ef9c49e. Migration status is migrating {{(pid=62692) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 887.591231] env[62692]: INFO nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Updating resource usage from migration 82adcbf0-fa69-4f63-b376-ff031ef9c49e [ 887.619818] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance d99737d8-2eb0-40ee-b61e-6c736c84ea59 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 887.620852] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 628cc505-3edf-4066-91be-da009ebcf219 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 887.620852] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance abee88ff-5cf5-4bf1-91e0-93b19cf30046 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 887.620852] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 5a08fa8a-f9fe-4879-bb7b-baa04097df6c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 887.620852] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 9b8e830c-61b7-4dd2-8324-d3a96eec1465 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 887.620852] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance bb547773-d176-4c8e-a0fa-a374d5050b1b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 887.620852] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 887.621098] env[62692]: WARNING nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 27583cb4-0478-44d8-8970-478682271513 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 887.621185] env[62692]: WARNING nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance c6a965d3-1c52-4f95-a226-9d15b7197ce7 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 887.621313] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 64a8329d-e0e9-4c2d-bd1f-844ee40eb980 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 887.621481] env[62692]: WARNING nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 18eb004f-a16b-477a-8d05-ed23a418b069 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 887.621621] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance e2383a6a-3581-40fc-a0eb-6981acdbf54a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 887.621788] env[62692]: WARNING nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 6b2d3385-a9af-432b-a1e7-c91adb2e1f69 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 887.621947] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 754b3c65-1e4b-49d2-8980-095d975edb01 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 887.622363] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 26c1ff5b-9eda-4131-a4ea-d4511a2364b7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 887.622506] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance f44e2ee2-6321-4237-9644-a19b5e90ae16 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 887.622716] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 433f740c-bde4-4d33-baca-5a7ac8179da1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 887.622854] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance b389b14d-6dcc-4b26-84f8-38b952bf6ecd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 887.760555] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141265, 'name': CreateVM_Task, 'duration_secs': 0.496296} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.760555] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 887.760555] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.760555] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.761019] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 887.761314] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ffc927c-0e2e-4faa-9706-05b4d7989b18 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.766261] env[62692]: DEBUG oslo_vmware.api [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Waiting for the task: (returnval){ [ 887.766261] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5214197f-4635-2a25-83b7-d9393dcbf176" [ 887.766261] env[62692]: _type = "Task" [ 887.766261] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.776247] env[62692]: DEBUG oslo_vmware.api [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5214197f-4635-2a25-83b7-d9393dcbf176, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.928203] env[62692]: DEBUG nova.compute.manager [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 888.065226] env[62692]: INFO nova.compute.manager [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Rebuilding instance [ 888.109442] env[62692]: DEBUG nova.compute.manager [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 888.110613] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee43036f-264b-41f2-8cf3-30d5caad797c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.129205] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 888.129465] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Migration 82adcbf0-fa69-4f63-b376-ff031ef9c49e is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 888.129655] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 2ccb76cf-f641-4306-a137-fb2417285df9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.287022] env[62692]: DEBUG oslo_vmware.api [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5214197f-4635-2a25-83b7-d9393dcbf176, 'name': SearchDatastore_Task, 'duration_secs': 0.023484} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.288674] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.288931] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 888.289271] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.289557] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.289753] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 888.290064] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0588436d-52e7-446b-9de4-af4b59122237 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.307340] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 888.307340] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 888.307485] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95a7b462-b95f-43b4-9c53-61e89745c7ce {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.314464] env[62692]: DEBUG oslo_vmware.api [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Waiting for the task: (returnval){ [ 888.314464] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52bd7e9b-e3e8-d00f-4868-8b58b92b7899" [ 888.314464] env[62692]: _type = "Task" [ 888.314464] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.322839] env[62692]: DEBUG oslo_vmware.api [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52bd7e9b-e3e8-d00f-4868-8b58b92b7899, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.397424] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07f12e43-2ba7-40e8-8141-51ce68e1a9a3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.417509] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Updating instance '2ccb76cf-f641-4306-a137-fb2417285df9' progress to 0 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 888.446103] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.623608] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 888.624092] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8165c9f9-12bf-42fb-a145-3042257583a6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.632348] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance ae44f074-a5c8-4259-99c6-9ce290b1570b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 888.634977] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 888.634977] env[62692]: value = "task-1141266" [ 888.634977] env[62692]: _type = "Task" [ 888.634977] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.645658] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141266, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.826785] env[62692]: DEBUG oslo_vmware.api [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52bd7e9b-e3e8-d00f-4868-8b58b92b7899, 'name': SearchDatastore_Task, 'duration_secs': 0.051445} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.827697] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7010f5dd-4e80-40e1-bb88-6a5a9be25937 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.834546] env[62692]: DEBUG oslo_vmware.api [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Waiting for the task: (returnval){ [ 888.834546] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52b9a63c-679a-a661-2df2-a183bd2be386" [ 888.834546] env[62692]: _type = "Task" [ 888.834546] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.844629] env[62692]: DEBUG oslo_vmware.api [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52b9a63c-679a-a661-2df2-a183bd2be386, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.923416] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 888.923736] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3c359c6e-0f47-4252-bfa3-d581a8052ddc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.931997] env[62692]: DEBUG oslo_vmware.api [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Waiting for the task: (returnval){ [ 888.931997] env[62692]: value = "task-1141267" [ 888.931997] env[62692]: _type = "Task" [ 888.931997] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.941772] env[62692]: DEBUG oslo_vmware.api [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141267, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.136909] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 889.149708] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141266, 'name': PowerOffVM_Task, 'duration_secs': 0.214115} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.150311] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 889.150311] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 889.151104] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-400ad8fc-56f9-4958-b324-245ad6649026 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.159179] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 889.159453] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c75efc93-a309-4318-af9e-0681473affe8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.224956] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 889.224956] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 889.224956] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Deleting the datastore file [datastore2] 433f740c-bde4-4d33-baca-5a7ac8179da1 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 889.224956] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0f5ee432-5219-407e-a5d0-f032ccb842e5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.232494] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 889.232494] env[62692]: value = "task-1141269" [ 889.232494] env[62692]: _type = "Task" [ 889.232494] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.241188] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141269, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.347102] env[62692]: DEBUG oslo_vmware.api [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52b9a63c-679a-a661-2df2-a183bd2be386, 'name': SearchDatastore_Task, 'duration_secs': 0.021444} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.347431] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.347682] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] b389b14d-6dcc-4b26-84f8-38b952bf6ecd/b389b14d-6dcc-4b26-84f8-38b952bf6ecd.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 889.347946] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d7183ecb-b5cd-49bf-9c92-c5247c0ddf6e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.356008] env[62692]: DEBUG oslo_vmware.api [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Waiting for the task: (returnval){ [ 889.356008] env[62692]: value = "task-1141270" [ 889.356008] env[62692]: _type = "Task" [ 889.356008] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.366242] env[62692]: DEBUG oslo_vmware.api [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Task: {'id': task-1141270, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.443465] env[62692]: DEBUG oslo_vmware.api [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141267, 'name': PowerOffVM_Task, 'duration_secs': 0.186465} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.444346] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 889.444518] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Updating instance '2ccb76cf-f641-4306-a137-fb2417285df9' progress to 17 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 889.644170] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 3fd2959b-80cb-470c-8c2a-40b7630458ed has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 889.745334] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141269, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170022} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.745636] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 889.745828] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 889.746017] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 889.868870] env[62692]: DEBUG oslo_vmware.api [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Task: {'id': task-1141270, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.509029} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.870326] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] b389b14d-6dcc-4b26-84f8-38b952bf6ecd/b389b14d-6dcc-4b26-84f8-38b952bf6ecd.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 889.870326] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 889.870326] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-29f64554-c404-4eba-af06-a702d9bd0fe6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.882347] env[62692]: DEBUG oslo_vmware.api [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Waiting for the task: (returnval){ [ 889.882347] env[62692]: value = "task-1141271" [ 889.882347] env[62692]: _type = "Task" [ 889.882347] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.895691] env[62692]: DEBUG oslo_vmware.api [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Task: {'id': task-1141271, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.958417] env[62692]: DEBUG nova.virt.hardware [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 889.958417] env[62692]: DEBUG nova.virt.hardware [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 889.958417] env[62692]: DEBUG nova.virt.hardware [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 889.958417] env[62692]: DEBUG nova.virt.hardware [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 889.958417] env[62692]: DEBUG nova.virt.hardware [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 889.958417] env[62692]: DEBUG nova.virt.hardware [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 889.958417] env[62692]: DEBUG nova.virt.hardware [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 889.959149] env[62692]: DEBUG nova.virt.hardware [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 889.959712] env[62692]: DEBUG nova.virt.hardware [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 889.960619] env[62692]: DEBUG nova.virt.hardware [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 889.960941] env[62692]: DEBUG nova.virt.hardware [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 889.966971] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ea5b5c4a-8661-42a8-9c09-5ab30f314349 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.989431] env[62692]: DEBUG oslo_vmware.api [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Waiting for the task: (returnval){ [ 889.989431] env[62692]: value = "task-1141272" [ 889.989431] env[62692]: _type = "Task" [ 889.989431] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.997247] env[62692]: DEBUG oslo_vmware.api [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141272, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.151401] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 44699df6-5262-4700-a096-8c0e450323d2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 890.395115] env[62692]: DEBUG oslo_vmware.api [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Task: {'id': task-1141271, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.092953} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.396241] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 890.398312] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ff41d26-5599-4a61-a7be-4eca0ced352b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.426480] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] b389b14d-6dcc-4b26-84f8-38b952bf6ecd/b389b14d-6dcc-4b26-84f8-38b952bf6ecd.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 890.427243] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53a7b8a6-6406-4610-a610-9443cdc5d9ba {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.452238] env[62692]: DEBUG oslo_vmware.api [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Waiting for the task: (returnval){ [ 890.452238] env[62692]: value = "task-1141273" [ 890.452238] env[62692]: _type = "Task" [ 890.452238] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.463378] env[62692]: DEBUG oslo_vmware.api [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Task: {'id': task-1141273, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.497369] env[62692]: DEBUG oslo_vmware.api [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141272, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.653239] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 890.798327] env[62692]: DEBUG nova.virt.hardware [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 890.798327] env[62692]: DEBUG nova.virt.hardware [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 890.798327] env[62692]: DEBUG nova.virt.hardware [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 890.798327] env[62692]: DEBUG nova.virt.hardware [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 890.802964] env[62692]: DEBUG nova.virt.hardware [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 890.802964] env[62692]: DEBUG nova.virt.hardware [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 890.804652] env[62692]: DEBUG nova.virt.hardware [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 890.804652] env[62692]: DEBUG nova.virt.hardware [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 890.804652] env[62692]: DEBUG nova.virt.hardware [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 890.804652] env[62692]: DEBUG nova.virt.hardware [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 890.804652] env[62692]: DEBUG nova.virt.hardware [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 890.805300] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cb9a01a-715e-4767-9f5c-bc3c7b3898e3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.815537] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c47dad30-a102-406a-8d59-fa7a577e2937 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.830024] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:55:28:27', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c842425c-544e-4ce2-9657-512723bd318e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5cd6131b-792a-47dc-afcc-399d6be018ae', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 890.837374] env[62692]: DEBUG oslo.service.loopingcall [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.837583] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 890.837799] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-95cd1d37-0871-4291-87e7-61a29541aadd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.860701] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 890.860701] env[62692]: value = "task-1141274" [ 890.860701] env[62692]: _type = "Task" [ 890.860701] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.868362] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141274, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.962957] env[62692]: DEBUG oslo_vmware.api [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Task: {'id': task-1141273, 'name': ReconfigVM_Task, 'duration_secs': 0.309982} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.963120] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Reconfigured VM instance instance-0000003f to attach disk [datastore2] b389b14d-6dcc-4b26-84f8-38b952bf6ecd/b389b14d-6dcc-4b26-84f8-38b952bf6ecd.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 890.964025] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2a7bdac2-a8cb-4fb5-bf1e-c389f9b9c5e6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.971150] env[62692]: DEBUG oslo_vmware.api [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Waiting for the task: (returnval){ [ 890.971150] env[62692]: value = "task-1141275" [ 890.971150] env[62692]: _type = "Task" [ 890.971150] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.979903] env[62692]: DEBUG oslo_vmware.api [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Task: {'id': task-1141275, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.998390] env[62692]: DEBUG oslo_vmware.api [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141272, 'name': ReconfigVM_Task, 'duration_secs': 0.599006} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.998738] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Updating instance '2ccb76cf-f641-4306-a137-fb2417285df9' progress to 33 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 891.156732] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance fac426b6-747a-455e-85a1-44e71b432180 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 891.370343] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141274, 'name': CreateVM_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.482543] env[62692]: DEBUG oslo_vmware.api [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Task: {'id': task-1141275, 'name': Rename_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.508247] env[62692]: DEBUG nova.virt.hardware [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:32:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='a53d9694-cbf0-4c37-a7c0-aa52b2363bc3',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1692421942',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 891.508247] env[62692]: DEBUG nova.virt.hardware [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 891.508247] env[62692]: DEBUG nova.virt.hardware [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 891.508247] env[62692]: DEBUG nova.virt.hardware [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 891.508247] env[62692]: DEBUG nova.virt.hardware [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 891.508247] env[62692]: DEBUG nova.virt.hardware [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 891.508247] env[62692]: DEBUG nova.virt.hardware [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 891.508247] env[62692]: DEBUG nova.virt.hardware [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 891.508247] env[62692]: DEBUG nova.virt.hardware [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 891.508515] env[62692]: DEBUG nova.virt.hardware [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 891.508515] env[62692]: DEBUG nova.virt.hardware [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 891.513813] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Reconfiguring VM instance instance-00000035 to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 891.514521] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f26cfa8d-9dc0-4330-a53c-3083252dd795 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.535353] env[62692]: DEBUG oslo_vmware.api [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Waiting for the task: (returnval){ [ 891.535353] env[62692]: value = "task-1141276" [ 891.535353] env[62692]: _type = "Task" [ 891.535353] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.546412] env[62692]: DEBUG oslo_vmware.api [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141276, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.660172] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 891.870723] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141274, 'name': CreateVM_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.988724] env[62692]: DEBUG oslo_vmware.api [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Task: {'id': task-1141275, 'name': Rename_Task, 'duration_secs': 0.887045} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.989681] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 891.990749] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-541de45e-81ea-4de1-bca8-d647011bbaa7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.003461] env[62692]: DEBUG oslo_vmware.api [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Waiting for the task: (returnval){ [ 892.003461] env[62692]: value = "task-1141277" [ 892.003461] env[62692]: _type = "Task" [ 892.003461] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.010819] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Acquiring lock "e2b3b062-10e2-4a9c-b6b3-f0786de07b79" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.011303] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Lock "e2b3b062-10e2-4a9c-b6b3-f0786de07b79" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.015597] env[62692]: DEBUG oslo_vmware.api [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Task: {'id': task-1141277, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.051716] env[62692]: DEBUG oslo_vmware.api [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141276, 'name': ReconfigVM_Task, 'duration_secs': 0.165826} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.052289] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Reconfigured VM instance instance-00000035 to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 892.053238] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a64aa43-1606-4ca9-8b1a-d6998d260845 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.085784] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] 2ccb76cf-f641-4306-a137-fb2417285df9/2ccb76cf-f641-4306-a137-fb2417285df9.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 892.086647] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-22ff9019-0b58-432d-9ac1-046f9a79b23a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.108551] env[62692]: DEBUG oslo_vmware.api [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Waiting for the task: (returnval){ [ 892.108551] env[62692]: value = "task-1141278" [ 892.108551] env[62692]: _type = "Task" [ 892.108551] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.116715] env[62692]: DEBUG oslo_vmware.api [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141278, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.163291] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance b23d3484-4fc5-46b2-8e8c-bdaab087f046 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 892.370837] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141274, 'name': CreateVM_Task, 'duration_secs': 1.152446} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.371034] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 892.371753] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.371926] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.372311] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 892.372596] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dee53711-225b-40b8-ada7-27d06a1311bd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.379647] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 892.379647] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52fe1df4-5ee9-c26a-5dea-65605f8700de" [ 892.379647] env[62692]: _type = "Task" [ 892.379647] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.389699] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52fe1df4-5ee9-c26a-5dea-65605f8700de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.512983] env[62692]: DEBUG oslo_vmware.api [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Task: {'id': task-1141277, 'name': PowerOnVM_Task, 'duration_secs': 0.469588} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.513293] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 892.513501] env[62692]: INFO nova.compute.manager [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Took 8.51 seconds to spawn the instance on the hypervisor. [ 892.513704] env[62692]: DEBUG nova.compute.manager [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 892.514486] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c94485ed-f8d2-40c5-a2a4-73eac599c365 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.623023] env[62692]: DEBUG oslo_vmware.api [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141278, 'name': ReconfigVM_Task, 'duration_secs': 0.301083} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.623023] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Reconfigured VM instance instance-00000035 to attach disk [datastore1] 2ccb76cf-f641-4306-a137-fb2417285df9/2ccb76cf-f641-4306-a137-fb2417285df9.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 892.623023] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Updating instance '2ccb76cf-f641-4306-a137-fb2417285df9' progress to 50 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 892.667178] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 8a56e6df-eea1-41f4-9360-4f06d2f516a3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 892.667178] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Total usable vcpus: 48, total allocated vcpus: 16 {{(pid=62692) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 892.667178] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3648MB phys_disk=200GB used_disk=15GB total_vcpus=48 used_vcpus=16 pci_stats=[] {{(pid=62692) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 892.895602] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52fe1df4-5ee9-c26a-5dea-65605f8700de, 'name': SearchDatastore_Task, 'duration_secs': 0.010369} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.898642] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.898943] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 892.899321] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.899548] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.899777] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 892.900563] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ff514348-4895-41bb-98e5-4bd850a3f950 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.911076] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 892.911336] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 892.914688] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39656332-fcfc-4dbc-90eb-a6e859a66387 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.922034] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 892.922034] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]522aaff0-71a8-75f0-5e29-bdd359301e15" [ 892.922034] env[62692]: _type = "Task" [ 892.922034] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.933228] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]522aaff0-71a8-75f0-5e29-bdd359301e15, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.035158] env[62692]: INFO nova.compute.manager [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Took 43.12 seconds to build instance. [ 893.087124] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d272109-339f-42e8-b7db-e6a8cc0234dd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.096283] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75e9720c-c942-4a49-899d-653657087599 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.127850] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c283419a-789a-4d6b-8df1-34c484d9e5e8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.132597] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42c04a5d-52f3-4773-ba7d-0289ec9237a5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.155527] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d7ae381-d3af-4447-9fdb-e4950996c4e5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.160218] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6a3be37-62ed-4d93-83b0-f35f76c7bbf5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.175509] env[62692]: DEBUG nova.compute.provider_tree [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 893.189729] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Updating instance '2ccb76cf-f641-4306-a137-fb2417285df9' progress to 67 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 893.215420] env[62692]: ERROR nova.scheduler.client.report [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [req-3c8989d6-84b5-48ec-9258-d35619608d91] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 470758a8-5efa-4875-aad5-f512a727752f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-3c8989d6-84b5-48ec-9258-d35619608d91"}]} [ 893.231065] env[62692]: DEBUG nova.scheduler.client.report [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Refreshing inventories for resource provider 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 893.250473] env[62692]: DEBUG nova.scheduler.client.report [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Updating ProviderTree inventory for provider 470758a8-5efa-4875-aad5-f512a727752f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 893.250670] env[62692]: DEBUG nova.compute.provider_tree [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 893.261472] env[62692]: DEBUG nova.scheduler.client.report [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Refreshing aggregate associations for resource provider 470758a8-5efa-4875-aad5-f512a727752f, aggregates: None {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 893.278549] env[62692]: DEBUG nova.scheduler.client.report [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Refreshing trait associations for resource provider 470758a8-5efa-4875-aad5-f512a727752f, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 893.435426] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]522aaff0-71a8-75f0-5e29-bdd359301e15, 'name': SearchDatastore_Task, 'duration_secs': 0.010129} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.436264] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e8c87af-3293-4fd5-9d65-ad5ea0194369 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.445214] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 893.445214] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52648ead-c1ef-22a9-acfb-e5509b3209c1" [ 893.445214] env[62692]: _type = "Task" [ 893.445214] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.455021] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52648ead-c1ef-22a9-acfb-e5509b3209c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.536967] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c3940dea-8fef-4707-bcbd-62f4d180639b tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Lock "b389b14d-6dcc-4b26-84f8-38b952bf6ecd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.232s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.660026] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-166bfd82-c5a2-46fd-a4f3-565a1704b7f4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.668124] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b4eb5c7-dfde-40c3-85cd-7257097bb82c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.701642] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64090132-fc47-47b6-8a66-432387cafb6a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.709931] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb994ba7-9b38-44b7-bb53-37bb5c4c08c9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.725317] env[62692]: DEBUG nova.compute.provider_tree [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 893.749810] env[62692]: DEBUG nova.network.neutron [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Port 951b0add-cfaa-4506-a6ac-56bcb8d1041e binding to destination host cpu-1 is already ACTIVE {{(pid=62692) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 893.956457] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52648ead-c1ef-22a9-acfb-e5509b3209c1, 'name': SearchDatastore_Task, 'duration_secs': 0.010116} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.956742] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.957068] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 433f740c-bde4-4d33-baca-5a7ac8179da1/433f740c-bde4-4d33-baca-5a7ac8179da1.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 893.957351] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-31cbeb9c-0c9d-4497-a5f5-a3dbd81bb147 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.965535] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 893.965535] env[62692]: value = "task-1141279" [ 893.965535] env[62692]: _type = "Task" [ 893.965535] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.974956] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141279, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.039905] env[62692]: DEBUG nova.compute.manager [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 894.267280] env[62692]: DEBUG nova.scheduler.client.report [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Updated inventory for provider 470758a8-5efa-4875-aad5-f512a727752f with generation 88 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 894.267473] env[62692]: DEBUG nova.compute.provider_tree [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Updating resource provider 470758a8-5efa-4875-aad5-f512a727752f generation from 88 to 89 during operation: update_inventory {{(pid=62692) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 894.267626] env[62692]: DEBUG nova.compute.provider_tree [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 894.304969] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Acquiring lock "b389b14d-6dcc-4b26-84f8-38b952bf6ecd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.305474] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Lock "b389b14d-6dcc-4b26-84f8-38b952bf6ecd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.305764] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Acquiring lock "b389b14d-6dcc-4b26-84f8-38b952bf6ecd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.305983] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Lock "b389b14d-6dcc-4b26-84f8-38b952bf6ecd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.306180] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Lock "b389b14d-6dcc-4b26-84f8-38b952bf6ecd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.309495] env[62692]: INFO nova.compute.manager [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Terminating instance [ 894.311948] env[62692]: DEBUG nova.compute.manager [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 894.312091] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 894.313182] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d5132aa-d240-461d-be76-a81e68eec92d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.323129] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 894.323430] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7ae9d623-3d22-4c03-94f8-4a27b8439451 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.332056] env[62692]: DEBUG oslo_vmware.api [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Waiting for the task: (returnval){ [ 894.332056] env[62692]: value = "task-1141280" [ 894.332056] env[62692]: _type = "Task" [ 894.332056] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.342330] env[62692]: DEBUG oslo_vmware.api [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Task: {'id': task-1141280, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.476230] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141279, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.506353} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.476496] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 433f740c-bde4-4d33-baca-5a7ac8179da1/433f740c-bde4-4d33-baca-5a7ac8179da1.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 894.477390] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 894.477390] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-18084a85-5a48-4650-8202-aa4cb79bb4f2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.485735] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 894.485735] env[62692]: value = "task-1141281" [ 894.485735] env[62692]: _type = "Task" [ 894.485735] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.499465] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141281, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.561351] env[62692]: DEBUG oslo_concurrency.lockutils [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.770898] env[62692]: DEBUG oslo_concurrency.lockutils [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquiring lock "2ccb76cf-f641-4306-a137-fb2417285df9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.771202] env[62692]: DEBUG oslo_concurrency.lockutils [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Lock "2ccb76cf-f641-4306-a137-fb2417285df9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.771335] env[62692]: DEBUG oslo_concurrency.lockutils [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Lock "2ccb76cf-f641-4306-a137-fb2417285df9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.772955] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62692) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 894.773186] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 8.208s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.773629] env[62692]: DEBUG oslo_concurrency.lockutils [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 40.181s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.775106] env[62692]: INFO nova.compute.claims [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 894.777506] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 894.777650] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Cleaning up deleted instances {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 894.842188] env[62692]: DEBUG oslo_vmware.api [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Task: {'id': task-1141280, 'name': PowerOffVM_Task, 'duration_secs': 0.239929} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.842467] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 894.842639] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 894.843117] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f2986a4b-16dd-4fe0-80f2-589b1c25514f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.907164] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 894.907399] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 894.907600] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Deleting the datastore file [datastore2] b389b14d-6dcc-4b26-84f8-38b952bf6ecd {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 894.907875] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ad81f2c6-b3f7-49c2-92bf-09cb28e26180 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.914730] env[62692]: DEBUG oslo_vmware.api [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Waiting for the task: (returnval){ [ 894.914730] env[62692]: value = "task-1141283" [ 894.914730] env[62692]: _type = "Task" [ 894.914730] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.923508] env[62692]: DEBUG oslo_vmware.api [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Task: {'id': task-1141283, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.995807] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141281, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074289} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.996105] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 894.996898] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1491a5e-9b63-4f6d-b323-37ed3ffc1f9f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.020475] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] 433f740c-bde4-4d33-baca-5a7ac8179da1/433f740c-bde4-4d33-baca-5a7ac8179da1.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 895.020775] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-481a4f4e-1514-46a6-ac59-302a797e2fbf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.042816] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 895.042816] env[62692]: value = "task-1141284" [ 895.042816] env[62692]: _type = "Task" [ 895.042816] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.051530] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141284, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.285726] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] There are 12 instances to clean {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 895.285989] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 8b75e781-930e-4885-81d2-8a8929d6c39b] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 895.426821] env[62692]: DEBUG oslo_vmware.api [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Task: {'id': task-1141283, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.554862] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141284, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.789609] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: ae0f1e48-0f0c-406f-bc99-2e3b7b86c0ba] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 895.826479] env[62692]: DEBUG oslo_concurrency.lockutils [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquiring lock "refresh_cache-2ccb76cf-f641-4306-a137-fb2417285df9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.826679] env[62692]: DEBUG oslo_concurrency.lockutils [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquired lock "refresh_cache-2ccb76cf-f641-4306-a137-fb2417285df9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.826864] env[62692]: DEBUG nova.network.neutron [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 895.927883] env[62692]: DEBUG oslo_vmware.api [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Task: {'id': task-1141283, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.690641} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.928176] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 895.928373] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 895.928556] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 895.928736] env[62692]: INFO nova.compute.manager [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Took 1.62 seconds to destroy the instance on the hypervisor. [ 895.929015] env[62692]: DEBUG oslo.service.loopingcall [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.929227] env[62692]: DEBUG nova.compute.manager [-] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 895.929514] env[62692]: DEBUG nova.network.neutron [-] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 896.060230] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141284, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.200029] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a280c6e-4736-4593-872f-cee5a925dd7a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.208015] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59766439-b2d4-4ab4-be6b-c378303a92b3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.244082] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c3352b3-bda9-4a95-b172-25486a2ddd1a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.254721] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-210e539d-fc65-4d5d-b8f2-abeb3bd669cb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.261496] env[62692]: DEBUG nova.compute.manager [req-cac44a11-c5c7-4976-8cf1-7ea67a714ea7 req-83dacb73-ead5-450b-9327-b3b3228c5e95 service nova] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Received event network-vif-deleted-54f5eeb9-6d87-4308-94cd-8441cd77972b {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 896.261712] env[62692]: INFO nova.compute.manager [req-cac44a11-c5c7-4976-8cf1-7ea67a714ea7 req-83dacb73-ead5-450b-9327-b3b3228c5e95 service nova] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Neutron deleted interface 54f5eeb9-6d87-4308-94cd-8441cd77972b; detaching it from the instance and deleting it from the info cache [ 896.261892] env[62692]: DEBUG nova.network.neutron [req-cac44a11-c5c7-4976-8cf1-7ea67a714ea7 req-83dacb73-ead5-450b-9327-b3b3228c5e95 service nova] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.275497] env[62692]: DEBUG nova.compute.provider_tree [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 896.295522] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: b1d6409a-2733-470f-a929-672fe1631b1b] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 896.531758] env[62692]: DEBUG nova.network.neutron [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Updating instance_info_cache with network_info: [{"id": "951b0add-cfaa-4506-a6ac-56bcb8d1041e", "address": "fa:16:3e:70:90:85", "network": {"id": "a74c4d30-3568-4e50-98f5-d2fcc05ec728", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.89", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "989968e8796a4a70a4dbd045418a475f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap951b0add-cf", "ovs_interfaceid": "951b0add-cfaa-4506-a6ac-56bcb8d1041e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.555519] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141284, 'name': ReconfigVM_Task, 'duration_secs': 1.504023} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.555827] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Reconfigured VM instance instance-0000003e to attach disk [datastore1] 433f740c-bde4-4d33-baca-5a7ac8179da1/433f740c-bde4-4d33-baca-5a7ac8179da1.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 896.556566] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6f926a97-c0a8-4cdf-b7a2-57a523f60b2b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.564424] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 896.564424] env[62692]: value = "task-1141285" [ 896.564424] env[62692]: _type = "Task" [ 896.564424] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.575124] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141285, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.749125] env[62692]: DEBUG nova.network.neutron [-] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.764861] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4b54715d-0dd2-47ad-87fa-a52538b3f4e6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.775040] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b202684-fcf6-4067-8a22-9a8067b472ba {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.800041] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 7a3f9c7d-03cf-4177-88db-acf33cc03feb] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 896.814263] env[62692]: DEBUG nova.compute.manager [req-cac44a11-c5c7-4976-8cf1-7ea67a714ea7 req-83dacb73-ead5-450b-9327-b3b3228c5e95 service nova] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Detach interface failed, port_id=54f5eeb9-6d87-4308-94cd-8441cd77972b, reason: Instance b389b14d-6dcc-4b26-84f8-38b952bf6ecd could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 896.823730] env[62692]: DEBUG nova.scheduler.client.report [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Updated inventory for provider 470758a8-5efa-4875-aad5-f512a727752f with generation 89 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 896.824042] env[62692]: DEBUG nova.compute.provider_tree [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Updating resource provider 470758a8-5efa-4875-aad5-f512a727752f generation from 89 to 90 during operation: update_inventory {{(pid=62692) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 896.824264] env[62692]: DEBUG nova.compute.provider_tree [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 897.034477] env[62692]: DEBUG oslo_concurrency.lockutils [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Releasing lock "refresh_cache-2ccb76cf-f641-4306-a137-fb2417285df9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.075577] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141285, 'name': Rename_Task, 'duration_secs': 0.16768} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.076513] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 897.076513] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f8b0e67b-efd6-44ec-888d-3a5ea0ef4227 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.084430] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 897.084430] env[62692]: value = "task-1141286" [ 897.084430] env[62692]: _type = "Task" [ 897.084430] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.092859] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141286, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.251908] env[62692]: INFO nova.compute.manager [-] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Took 1.32 seconds to deallocate network for instance. [ 897.315043] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 1cdd1dbc-567d-46f6-9f05-dfbc57acbf99] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 897.329443] env[62692]: DEBUG oslo_concurrency.lockutils [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.556s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.330127] env[62692]: DEBUG nova.compute.manager [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 897.332633] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 41.419s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.334018] env[62692]: INFO nova.compute.claims [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 897.560665] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1e50c2c-5cca-47ae-93d9-e2fa04c0ebb3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.581827] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a4f3460-0a67-42f0-890e-7758d904ffc0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.593077] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Updating instance '2ccb76cf-f641-4306-a137-fb2417285df9' progress to 83 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 897.600572] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141286, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.759562] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.818734] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 1848c251-0a0c-4c36-8bd0-ff37befde2a0] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 897.840308] env[62692]: DEBUG nova.compute.utils [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 897.841610] env[62692]: DEBUG nova.compute.manager [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 897.841831] env[62692]: DEBUG nova.network.neutron [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 897.889907] env[62692]: DEBUG nova.policy [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd3c6698b721144c9adac1fd0c9c7a073', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7db5465cb5db465b92dfa570572b6e80', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 898.095051] env[62692]: DEBUG oslo_vmware.api [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141286, 'name': PowerOnVM_Task, 'duration_secs': 0.525305} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.095352] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 898.095561] env[62692]: DEBUG nova.compute.manager [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 898.096372] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf33287c-510c-45ec-9d9a-d115e1fb3733 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.104233] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 898.106813] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a57327e6-ef9e-4178-8c2d-239dcc9637a9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.114480] env[62692]: DEBUG oslo_vmware.api [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Waiting for the task: (returnval){ [ 898.114480] env[62692]: value = "task-1141287" [ 898.114480] env[62692]: _type = "Task" [ 898.114480] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.122300] env[62692]: DEBUG oslo_vmware.api [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141287, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.232840] env[62692]: DEBUG nova.network.neutron [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Successfully created port: 2c7422d2-dfef-4010-a99a-60c510c54a3c {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 898.321798] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: aa0666ce-4d4f-4333-8acb-ea0cfb438cd9] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 898.345043] env[62692]: DEBUG nova.compute.manager [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 898.621788] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.629204] env[62692]: DEBUG oslo_vmware.api [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141287, 'name': PowerOnVM_Task, 'duration_secs': 0.41197} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.629543] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 898.629776] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-98f30c6b-5398-4f52-ab5e-82ff975d1715 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Updating instance '2ccb76cf-f641-4306-a137-fb2417285df9' progress to 100 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 898.785104] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-498f95f1-ccf8-4a45-b810-9d696ea74534 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.793383] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaacefab-35f6-4d08-bf52-d451264b7afd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.826217] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 6097d6b2-52d5-4765-94c0-d5f3609464d8] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 898.829634] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7af73aae-0367-435e-83e5-e2c5de9b8aac {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.839871] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34d5d2db-d84e-4b7d-ad32-a3e46a6f1e3d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.857183] env[62692]: DEBUG nova.compute.provider_tree [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 899.329250] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 105e6fd5-4eff-4a0c-9a4b-e6deade781d5] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 899.360044] env[62692]: DEBUG nova.compute.manager [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 899.362884] env[62692]: DEBUG nova.scheduler.client.report [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 899.385450] env[62692]: DEBUG nova.virt.hardware [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 899.385734] env[62692]: DEBUG nova.virt.hardware [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 899.385900] env[62692]: DEBUG nova.virt.hardware [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 899.386106] env[62692]: DEBUG nova.virt.hardware [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 899.386266] env[62692]: DEBUG nova.virt.hardware [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 899.386421] env[62692]: DEBUG nova.virt.hardware [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 899.386656] env[62692]: DEBUG nova.virt.hardware [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 899.386936] env[62692]: DEBUG nova.virt.hardware [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 899.387157] env[62692]: DEBUG nova.virt.hardware [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 899.387363] env[62692]: DEBUG nova.virt.hardware [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 899.387551] env[62692]: DEBUG nova.virt.hardware [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 899.388651] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fb0bd2d-f371-46a4-b965-a0e7cbaf44e7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.397594] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58ce8e4b-0fa1-42a0-8616-762f27f42146 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.784353] env[62692]: DEBUG oslo_concurrency.lockutils [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "433f740c-bde4-4d33-baca-5a7ac8179da1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.784595] env[62692]: DEBUG oslo_concurrency.lockutils [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "433f740c-bde4-4d33-baca-5a7ac8179da1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.784826] env[62692]: DEBUG oslo_concurrency.lockutils [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "433f740c-bde4-4d33-baca-5a7ac8179da1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.785037] env[62692]: DEBUG oslo_concurrency.lockutils [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "433f740c-bde4-4d33-baca-5a7ac8179da1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.785237] env[62692]: DEBUG oslo_concurrency.lockutils [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "433f740c-bde4-4d33-baca-5a7ac8179da1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.790251] env[62692]: INFO nova.compute.manager [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Terminating instance [ 899.790251] env[62692]: DEBUG nova.compute.manager [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 899.790251] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 899.790414] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99aea51f-fd44-4290-8c9b-981bc4b1ad72 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.799334] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 899.799871] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-89351ca7-c59c-4697-ba0f-0a267245a2c4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.807196] env[62692]: DEBUG oslo_vmware.api [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 899.807196] env[62692]: value = "task-1141288" [ 899.807196] env[62692]: _type = "Task" [ 899.807196] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.817068] env[62692]: DEBUG oslo_vmware.api [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141288, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.834762] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: c1818ed2-9e27-4738-bc17-98832318cf61] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 899.867954] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.535s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.868592] env[62692]: DEBUG nova.compute.manager [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 899.872518] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 40.587s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.877568] env[62692]: INFO nova.compute.claims [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 899.964242] env[62692]: DEBUG nova.compute.manager [req-13b0d11c-afb1-453e-ac3b-c3055ca206e8 req-2a73c6ba-ecf1-4f9a-8d1f-fe8b4bf82c7b service nova] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Received event network-vif-plugged-2c7422d2-dfef-4010-a99a-60c510c54a3c {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 899.964604] env[62692]: DEBUG oslo_concurrency.lockutils [req-13b0d11c-afb1-453e-ac3b-c3055ca206e8 req-2a73c6ba-ecf1-4f9a-8d1f-fe8b4bf82c7b service nova] Acquiring lock "3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.964951] env[62692]: DEBUG oslo_concurrency.lockutils [req-13b0d11c-afb1-453e-ac3b-c3055ca206e8 req-2a73c6ba-ecf1-4f9a-8d1f-fe8b4bf82c7b service nova] Lock "3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.965255] env[62692]: DEBUG oslo_concurrency.lockutils [req-13b0d11c-afb1-453e-ac3b-c3055ca206e8 req-2a73c6ba-ecf1-4f9a-8d1f-fe8b4bf82c7b service nova] Lock "3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.965528] env[62692]: DEBUG nova.compute.manager [req-13b0d11c-afb1-453e-ac3b-c3055ca206e8 req-2a73c6ba-ecf1-4f9a-8d1f-fe8b4bf82c7b service nova] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] No waiting events found dispatching network-vif-plugged-2c7422d2-dfef-4010-a99a-60c510c54a3c {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 899.965834] env[62692]: WARNING nova.compute.manager [req-13b0d11c-afb1-453e-ac3b-c3055ca206e8 req-2a73c6ba-ecf1-4f9a-8d1f-fe8b4bf82c7b service nova] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Received unexpected event network-vif-plugged-2c7422d2-dfef-4010-a99a-60c510c54a3c for instance with vm_state building and task_state spawning. [ 900.119235] env[62692]: DEBUG nova.network.neutron [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Successfully updated port: 2c7422d2-dfef-4010-a99a-60c510c54a3c {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 900.260403] env[62692]: DEBUG oslo_concurrency.lockutils [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "3f6c9744-a6e3-43f3-8b6b-624ba178747d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.260711] env[62692]: DEBUG oslo_concurrency.lockutils [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "3f6c9744-a6e3-43f3-8b6b-624ba178747d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.316912] env[62692]: DEBUG oslo_vmware.api [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141288, 'name': PowerOffVM_Task, 'duration_secs': 0.203505} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.317220] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 900.317391] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 900.317611] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a6ff366f-b9d4-4f72-9e75-58fb89e9d2d3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.338097] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: fa7bb883-0065-4ecc-9eb5-8c5443344a2a] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 900.385187] env[62692]: DEBUG nova.compute.utils [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 900.386629] env[62692]: DEBUG nova.compute.manager [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 900.386808] env[62692]: DEBUG nova.network.neutron [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 900.393442] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 900.393607] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 900.393644] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Deleting the datastore file [datastore1] 433f740c-bde4-4d33-baca-5a7ac8179da1 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 900.394242] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9263057b-a7ba-4f22-b1b6-78e10e298bc0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.402865] env[62692]: DEBUG oslo_vmware.api [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 900.402865] env[62692]: value = "task-1141290" [ 900.402865] env[62692]: _type = "Task" [ 900.402865] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.413505] env[62692]: DEBUG oslo_vmware.api [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141290, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.442244] env[62692]: DEBUG nova.policy [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '17517e7cf7174a7cb3fa45381c91f99f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fc06bb980c584d2980ee5eb48e678772', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 900.624149] env[62692]: DEBUG oslo_concurrency.lockutils [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "refresh_cache-3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.624149] env[62692]: DEBUG oslo_concurrency.lockutils [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquired lock "refresh_cache-3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.624149] env[62692]: DEBUG nova.network.neutron [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 900.726208] env[62692]: DEBUG nova.network.neutron [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Successfully created port: f6f97326-a148-4785-a942-d2869ddffca3 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 900.841330] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 54f615e3-5c6e-44bb-aeb7-5620a639b55c] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 900.892786] env[62692]: DEBUG nova.compute.manager [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 900.913449] env[62692]: DEBUG oslo_vmware.api [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141290, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140892} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.913721] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 900.913972] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 900.915432] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 900.915432] env[62692]: INFO nova.compute.manager [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Took 1.12 seconds to destroy the instance on the hypervisor. [ 900.915432] env[62692]: DEBUG oslo.service.loopingcall [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 900.915432] env[62692]: DEBUG nova.compute.manager [-] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 900.915432] env[62692]: DEBUG nova.network.neutron [-] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 901.172900] env[62692]: DEBUG nova.network.neutron [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 901.347851] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 901.347851] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Cleaning up deleted instances with incomplete migration {{(pid=62692) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 901.385123] env[62692]: DEBUG nova.network.neutron [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Updating instance_info_cache with network_info: [{"id": "2c7422d2-dfef-4010-a99a-60c510c54a3c", "address": "fa:16:3e:71:16:ae", "network": {"id": "17e72c3b-603f-4305-96ac-89a9b4c9c7e9", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-123501993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7db5465cb5db465b92dfa570572b6e80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c7422d2-df", "ovs_interfaceid": "2c7422d2-dfef-4010-a99a-60c510c54a3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.388697] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eb9a2fb-e678-46b7-be0f-8bcefb4bcd6d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.398093] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fac95d4c-8d26-459e-babb-218eec7919a0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.433320] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4a6726e-be75-4398-822b-fc3db8aeeeca {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.442633] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e2ceb9a-1a21-45ec-b6c9-7806a66956dc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.460935] env[62692]: DEBUG nova.compute.provider_tree [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 901.466306] env[62692]: DEBUG nova.compute.manager [req-e9f49edd-17f4-464b-bbc4-3653f76fe4d1 req-a0540a24-a7b3-4fb3-b41b-8499a56e5103 service nova] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Received event network-vif-deleted-5cd6131b-792a-47dc-afcc-399d6be018ae {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 901.466502] env[62692]: INFO nova.compute.manager [req-e9f49edd-17f4-464b-bbc4-3653f76fe4d1 req-a0540a24-a7b3-4fb3-b41b-8499a56e5103 service nova] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Neutron deleted interface 5cd6131b-792a-47dc-afcc-399d6be018ae; detaching it from the instance and deleting it from the info cache [ 901.467078] env[62692]: DEBUG nova.network.neutron [req-e9f49edd-17f4-464b-bbc4-3653f76fe4d1 req-a0540a24-a7b3-4fb3-b41b-8499a56e5103 service nova] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.597565] env[62692]: DEBUG nova.network.neutron [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Port 951b0add-cfaa-4506-a6ac-56bcb8d1041e binding to destination host cpu-1 is already ACTIVE {{(pid=62692) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 901.597850] env[62692]: DEBUG oslo_concurrency.lockutils [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquiring lock "refresh_cache-2ccb76cf-f641-4306-a137-fb2417285df9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.598024] env[62692]: DEBUG oslo_concurrency.lockutils [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquired lock "refresh_cache-2ccb76cf-f641-4306-a137-fb2417285df9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.598204] env[62692]: DEBUG nova.network.neutron [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 901.752382] env[62692]: DEBUG nova.network.neutron [-] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.850944] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 901.890797] env[62692]: DEBUG oslo_concurrency.lockutils [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Releasing lock "refresh_cache-3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.891125] env[62692]: DEBUG nova.compute.manager [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Instance network_info: |[{"id": "2c7422d2-dfef-4010-a99a-60c510c54a3c", "address": "fa:16:3e:71:16:ae", "network": {"id": "17e72c3b-603f-4305-96ac-89a9b4c9c7e9", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-123501993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7db5465cb5db465b92dfa570572b6e80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c7422d2-df", "ovs_interfaceid": "2c7422d2-dfef-4010-a99a-60c510c54a3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 901.891623] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:71:16:ae', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '12d8eedb-97cb-4d3b-b364-42d7fd8b3c85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2c7422d2-dfef-4010-a99a-60c510c54a3c', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 901.899302] env[62692]: DEBUG oslo.service.loopingcall [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.899806] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 901.900176] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-98d7ecaa-770c-45ab-a69c-ff7293f4323d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.917664] env[62692]: DEBUG nova.compute.manager [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 901.927711] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 901.927711] env[62692]: value = "task-1141291" [ 901.927711] env[62692]: _type = "Task" [ 901.927711] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.940673] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141291, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.942648] env[62692]: DEBUG nova.virt.hardware [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 901.942923] env[62692]: DEBUG nova.virt.hardware [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 901.943161] env[62692]: DEBUG nova.virt.hardware [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 901.943416] env[62692]: DEBUG nova.virt.hardware [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 901.944161] env[62692]: DEBUG nova.virt.hardware [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 901.944161] env[62692]: DEBUG nova.virt.hardware [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 901.944161] env[62692]: DEBUG nova.virt.hardware [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 901.944504] env[62692]: DEBUG nova.virt.hardware [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 901.944504] env[62692]: DEBUG nova.virt.hardware [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 901.944723] env[62692]: DEBUG nova.virt.hardware [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 901.945163] env[62692]: DEBUG nova.virt.hardware [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 901.945844] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-268c11fb-e9f2-4487-9dbe-596eb4c513b7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.954383] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c55e8937-2bc0-4988-8355-02e2974b453b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.962523] env[62692]: DEBUG nova.scheduler.client.report [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 901.973698] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b5108e99-4d04-4d0f-b86d-283733f25cf2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.983573] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c80aca9-d5db-4f4a-b633-28b91a866365 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.996907] env[62692]: DEBUG nova.compute.manager [req-f40f2b2a-690c-40e1-bf70-f7731489a25c req-5414c97b-a241-417d-b3a3-ed9fdc954969 service nova] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Received event network-changed-2c7422d2-dfef-4010-a99a-60c510c54a3c {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 901.997048] env[62692]: DEBUG nova.compute.manager [req-f40f2b2a-690c-40e1-bf70-f7731489a25c req-5414c97b-a241-417d-b3a3-ed9fdc954969 service nova] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Refreshing instance network info cache due to event network-changed-2c7422d2-dfef-4010-a99a-60c510c54a3c. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 901.997291] env[62692]: DEBUG oslo_concurrency.lockutils [req-f40f2b2a-690c-40e1-bf70-f7731489a25c req-5414c97b-a241-417d-b3a3-ed9fdc954969 service nova] Acquiring lock "refresh_cache-3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.997449] env[62692]: DEBUG oslo_concurrency.lockutils [req-f40f2b2a-690c-40e1-bf70-f7731489a25c req-5414c97b-a241-417d-b3a3-ed9fdc954969 service nova] Acquired lock "refresh_cache-3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.997618] env[62692]: DEBUG nova.network.neutron [req-f40f2b2a-690c-40e1-bf70-f7731489a25c req-5414c97b-a241-417d-b3a3-ed9fdc954969 service nova] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Refreshing network info cache for port 2c7422d2-dfef-4010-a99a-60c510c54a3c {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 902.021077] env[62692]: DEBUG nova.compute.manager [req-e9f49edd-17f4-464b-bbc4-3653f76fe4d1 req-a0540a24-a7b3-4fb3-b41b-8499a56e5103 service nova] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Detach interface failed, port_id=5cd6131b-792a-47dc-afcc-399d6be018ae, reason: Instance 433f740c-bde4-4d33-baca-5a7ac8179da1 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 902.254611] env[62692]: INFO nova.compute.manager [-] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Took 1.34 seconds to deallocate network for instance. [ 902.438608] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141291, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.474692] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.602s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.475260] env[62692]: DEBUG nova.compute.manager [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 902.478038] env[62692]: DEBUG oslo_concurrency.lockutils [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.953s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.478255] env[62692]: DEBUG oslo_concurrency.lockutils [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.480245] env[62692]: INFO nova.compute.manager [None req-50c5fcda-6f07-4f6e-a8d2-11125a35e104 tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Successfully reverted task state from rebuilding on failure for instance. [ 902.485722] env[62692]: DEBUG oslo_concurrency.lockutils [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.466s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.485923] env[62692]: DEBUG oslo_concurrency.lockutils [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.488136] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.882s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.488339] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.490092] env[62692]: DEBUG oslo_concurrency.lockutils [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.338s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.491564] env[62692]: INFO nova.compute.claims [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 902.512695] env[62692]: DEBUG nova.network.neutron [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Updating instance_info_cache with network_info: [{"id": "951b0add-cfaa-4506-a6ac-56bcb8d1041e", "address": "fa:16:3e:70:90:85", "network": {"id": "a74c4d30-3568-4e50-98f5-d2fcc05ec728", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.89", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "989968e8796a4a70a4dbd045418a475f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap951b0add-cf", "ovs_interfaceid": "951b0add-cfaa-4506-a6ac-56bcb8d1041e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.519967] env[62692]: INFO nova.scheduler.client.report [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Deleted allocations for instance c6a965d3-1c52-4f95-a226-9d15b7197ce7 [ 902.533802] env[62692]: INFO nova.scheduler.client.report [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Deleted allocations for instance 18eb004f-a16b-477a-8d05-ed23a418b069 [ 902.623362] env[62692]: DEBUG nova.network.neutron [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Successfully updated port: f6f97326-a148-4785-a942-d2869ddffca3 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 902.727377] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "a668a569-a90e-4e59-8d2e-b0225745b500" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.727638] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "a668a569-a90e-4e59-8d2e-b0225745b500" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.760965] env[62692]: DEBUG oslo_concurrency.lockutils [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.939864] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141291, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.996598] env[62692]: DEBUG nova.compute.utils [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 903.002104] env[62692]: DEBUG nova.compute.manager [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 903.002104] env[62692]: DEBUG nova.network.neutron [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 903.015461] env[62692]: DEBUG oslo_concurrency.lockutils [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Releasing lock "refresh_cache-2ccb76cf-f641-4306-a137-fb2417285df9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.029795] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a8a3268b-b859-47bb-b77b-166ee03fcee0 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884 tempest-FloatingIPsAssociationNegativeTestJSON-1311117884-project-member] Lock "c6a965d3-1c52-4f95-a226-9d15b7197ce7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.535s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.039058] env[62692]: DEBUG nova.network.neutron [req-f40f2b2a-690c-40e1-bf70-f7731489a25c req-5414c97b-a241-417d-b3a3-ed9fdc954969 service nova] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Updated VIF entry in instance network info cache for port 2c7422d2-dfef-4010-a99a-60c510c54a3c. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 903.039449] env[62692]: DEBUG nova.network.neutron [req-f40f2b2a-690c-40e1-bf70-f7731489a25c req-5414c97b-a241-417d-b3a3-ed9fdc954969 service nova] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Updating instance_info_cache with network_info: [{"id": "2c7422d2-dfef-4010-a99a-60c510c54a3c", "address": "fa:16:3e:71:16:ae", "network": {"id": "17e72c3b-603f-4305-96ac-89a9b4c9c7e9", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-123501993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7db5465cb5db465b92dfa570572b6e80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c7422d2-df", "ovs_interfaceid": "2c7422d2-dfef-4010-a99a-60c510c54a3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.043415] env[62692]: DEBUG oslo_concurrency.lockutils [None req-daf28a62-ed53-41a2-8cac-c0ea4fb78888 tempest-ServersNegativeTestMultiTenantJSON-1171099673 tempest-ServersNegativeTestMultiTenantJSON-1171099673-project-member] Lock "18eb004f-a16b-477a-8d05-ed23a418b069" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.217s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.073975] env[62692]: DEBUG nova.policy [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '24e7aa820d314f77ace38aa3a564d449', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4fed583a514c4043a4b5772c73e21aff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 903.126183] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Acquiring lock "refresh_cache-ae44f074-a5c8-4259-99c6-9ce290b1570b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.126346] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Acquired lock "refresh_cache-ae44f074-a5c8-4259-99c6-9ce290b1570b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.126502] env[62692]: DEBUG nova.network.neutron [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 903.444243] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141291, 'name': CreateVM_Task, 'duration_secs': 1.297378} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.444243] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 903.444984] env[62692]: DEBUG oslo_concurrency.lockutils [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.445319] env[62692]: DEBUG oslo_concurrency.lockutils [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.445746] env[62692]: DEBUG oslo_concurrency.lockutils [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 903.446112] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc340b39-b281-4e57-bd78-b2c0fa36db92 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.451789] env[62692]: DEBUG oslo_vmware.api [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 903.451789] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52bb631d-35be-f1d8-1afc-9e01e77ab723" [ 903.451789] env[62692]: _type = "Task" [ 903.451789] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.462230] env[62692]: DEBUG oslo_vmware.api [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52bb631d-35be-f1d8-1afc-9e01e77ab723, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.503686] env[62692]: DEBUG nova.compute.manager [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 903.518404] env[62692]: DEBUG nova.compute.manager [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62692) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 903.518548] env[62692]: DEBUG oslo_concurrency.lockutils [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.545995] env[62692]: DEBUG oslo_concurrency.lockutils [req-f40f2b2a-690c-40e1-bf70-f7731489a25c req-5414c97b-a241-417d-b3a3-ed9fdc954969 service nova] Releasing lock "refresh_cache-3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.628240] env[62692]: DEBUG nova.network.neutron [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Successfully created port: bfd79ef9-4a10-444c-96eb-1a4f4273a5d7 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 903.703622] env[62692]: DEBUG nova.network.neutron [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 903.968942] env[62692]: DEBUG oslo_vmware.api [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52bb631d-35be-f1d8-1afc-9e01e77ab723, 'name': SearchDatastore_Task, 'duration_secs': 0.010534} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.968942] env[62692]: DEBUG oslo_concurrency.lockutils [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.968942] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 903.968942] env[62692]: DEBUG oslo_concurrency.lockutils [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.968942] env[62692]: DEBUG oslo_concurrency.lockutils [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.968942] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 903.969826] env[62692]: DEBUG nova.network.neutron [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Updating instance_info_cache with network_info: [{"id": "f6f97326-a148-4785-a942-d2869ddffca3", "address": "fa:16:3e:9e:b5:42", "network": {"id": "14d97ab8-1b79-4ef2-a354-6398d00bdeba", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-707295669-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fc06bb980c584d2980ee5eb48e678772", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6f97326-a1", "ovs_interfaceid": "f6f97326-a148-4785-a942-d2869ddffca3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.970974] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-797445e4-09cf-4548-9ada-34d6368fe22d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.984028] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 903.984222] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 903.985569] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c05d86d8-9f7a-4073-95f1-5e64c08fab13 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.992675] env[62692]: DEBUG oslo_vmware.api [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 903.992675] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52300381-1b33-04f5-5b4c-79b00268c199" [ 903.992675] env[62692]: _type = "Task" [ 903.992675] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.994632] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1006af79-a653-4a3d-b917-0d5b0c0583ba {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.014433] env[62692]: DEBUG oslo_vmware.api [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52300381-1b33-04f5-5b4c-79b00268c199, 'name': SearchDatastore_Task, 'duration_secs': 0.010852} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.016069] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f2db5e6-5eb2-4a2a-8dad-2d27bedc524a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.019383] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-339cc37b-fe40-45f3-83a0-7cc900dd54dc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.025248] env[62692]: DEBUG nova.compute.manager [req-b7157021-eda6-4072-a041-7693d1184e0c req-08a174bd-a288-4c6b-b8fb-905369203573 service nova] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Received event network-vif-plugged-f6f97326-a148-4785-a942-d2869ddffca3 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 904.025248] env[62692]: DEBUG oslo_concurrency.lockutils [req-b7157021-eda6-4072-a041-7693d1184e0c req-08a174bd-a288-4c6b-b8fb-905369203573 service nova] Acquiring lock "ae44f074-a5c8-4259-99c6-9ce290b1570b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.025248] env[62692]: DEBUG oslo_concurrency.lockutils [req-b7157021-eda6-4072-a041-7693d1184e0c req-08a174bd-a288-4c6b-b8fb-905369203573 service nova] Lock "ae44f074-a5c8-4259-99c6-9ce290b1570b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.025475] env[62692]: DEBUG oslo_concurrency.lockutils [req-b7157021-eda6-4072-a041-7693d1184e0c req-08a174bd-a288-4c6b-b8fb-905369203573 service nova] Lock "ae44f074-a5c8-4259-99c6-9ce290b1570b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.025663] env[62692]: DEBUG nova.compute.manager [req-b7157021-eda6-4072-a041-7693d1184e0c req-08a174bd-a288-4c6b-b8fb-905369203573 service nova] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] No waiting events found dispatching network-vif-plugged-f6f97326-a148-4785-a942-d2869ddffca3 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 904.025857] env[62692]: WARNING nova.compute.manager [req-b7157021-eda6-4072-a041-7693d1184e0c req-08a174bd-a288-4c6b-b8fb-905369203573 service nova] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Received unexpected event network-vif-plugged-f6f97326-a148-4785-a942-d2869ddffca3 for instance with vm_state building and task_state spawning. [ 904.026095] env[62692]: DEBUG nova.compute.manager [req-b7157021-eda6-4072-a041-7693d1184e0c req-08a174bd-a288-4c6b-b8fb-905369203573 service nova] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Received event network-changed-f6f97326-a148-4785-a942-d2869ddffca3 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 904.027097] env[62692]: DEBUG nova.compute.manager [req-b7157021-eda6-4072-a041-7693d1184e0c req-08a174bd-a288-4c6b-b8fb-905369203573 service nova] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Refreshing instance network info cache due to event network-changed-f6f97326-a148-4785-a942-d2869ddffca3. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 904.027097] env[62692]: DEBUG oslo_concurrency.lockutils [req-b7157021-eda6-4072-a041-7693d1184e0c req-08a174bd-a288-4c6b-b8fb-905369203573 service nova] Acquiring lock "refresh_cache-ae44f074-a5c8-4259-99c6-9ce290b1570b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.057155] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0019e8e-03d9-4511-89e7-4d26cac50bf6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.061916] env[62692]: DEBUG oslo_vmware.api [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 904.061916] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52b638c5-d901-1587-870a-5def5a6fbaec" [ 904.061916] env[62692]: _type = "Task" [ 904.061916] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.070325] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c6fa09c-35bd-4ea6-863d-0d1f3a825953 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.078500] env[62692]: DEBUG oslo_vmware.api [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52b638c5-d901-1587-870a-5def5a6fbaec, 'name': SearchDatastore_Task, 'duration_secs': 0.011429} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.079889] env[62692]: DEBUG oslo_concurrency.lockutils [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.079889] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19/3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 904.079889] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f1e62f91-430a-4fd8-a31d-f2ba2c405df6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.090476] env[62692]: DEBUG nova.compute.provider_tree [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 904.097814] env[62692]: DEBUG oslo_vmware.api [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 904.097814] env[62692]: value = "task-1141292" [ 904.097814] env[62692]: _type = "Task" [ 904.097814] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.108638] env[62692]: DEBUG oslo_vmware.api [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141292, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.254940] env[62692]: DEBUG nova.network.neutron [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Successfully created port: b81f3e56-2a23-40b7-8153-e53be84131ca {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 904.477236] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Releasing lock "refresh_cache-ae44f074-a5c8-4259-99c6-9ce290b1570b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.477594] env[62692]: DEBUG nova.compute.manager [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Instance network_info: |[{"id": "f6f97326-a148-4785-a942-d2869ddffca3", "address": "fa:16:3e:9e:b5:42", "network": {"id": "14d97ab8-1b79-4ef2-a354-6398d00bdeba", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-707295669-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fc06bb980c584d2980ee5eb48e678772", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6f97326-a1", "ovs_interfaceid": "f6f97326-a148-4785-a942-d2869ddffca3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 904.478036] env[62692]: DEBUG oslo_concurrency.lockutils [req-b7157021-eda6-4072-a041-7693d1184e0c req-08a174bd-a288-4c6b-b8fb-905369203573 service nova] Acquired lock "refresh_cache-ae44f074-a5c8-4259-99c6-9ce290b1570b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.478359] env[62692]: DEBUG nova.network.neutron [req-b7157021-eda6-4072-a041-7693d1184e0c req-08a174bd-a288-4c6b-b8fb-905369203573 service nova] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Refreshing network info cache for port f6f97326-a148-4785-a942-d2869ddffca3 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 904.483711] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9e:b5:42', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60bdba1a-14cf-46b2-9d8b-aeaf4d80c815', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f6f97326-a148-4785-a942-d2869ddffca3', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 904.495051] env[62692]: DEBUG oslo.service.loopingcall [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 904.495432] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 904.495917] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ed576c49-7a11-4d94-9163-07f83a867ec3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.515917] env[62692]: DEBUG nova.compute.manager [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 904.521778] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 904.521778] env[62692]: value = "task-1141293" [ 904.521778] env[62692]: _type = "Task" [ 904.521778] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.538878] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141293, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.556194] env[62692]: DEBUG nova.virt.hardware [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 904.556552] env[62692]: DEBUG nova.virt.hardware [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 904.556770] env[62692]: DEBUG nova.virt.hardware [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 904.557052] env[62692]: DEBUG nova.virt.hardware [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 904.557268] env[62692]: DEBUG nova.virt.hardware [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 904.557463] env[62692]: DEBUG nova.virt.hardware [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 904.557726] env[62692]: DEBUG nova.virt.hardware [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 904.557957] env[62692]: DEBUG nova.virt.hardware [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 904.558200] env[62692]: DEBUG nova.virt.hardware [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 904.558390] env[62692]: DEBUG nova.virt.hardware [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 904.558575] env[62692]: DEBUG nova.virt.hardware [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 904.559863] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85d6f7f2-33f2-4e14-b0a3-350e2961c80a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.568853] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2facb64-691d-4f17-b49e-7ad2576e0545 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.593390] env[62692]: DEBUG nova.scheduler.client.report [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 904.609805] env[62692]: DEBUG oslo_vmware.api [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141292, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457098} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.609805] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19/3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 904.610088] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 904.610395] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-efd28d84-f62a-4b0b-b1d4-73bfe37ac3ec {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.619823] env[62692]: DEBUG oslo_vmware.api [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 904.619823] env[62692]: value = "task-1141294" [ 904.619823] env[62692]: _type = "Task" [ 904.619823] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.633766] env[62692]: DEBUG oslo_vmware.api [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141294, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.033923] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141293, 'name': CreateVM_Task, 'duration_secs': 0.332425} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.034624] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 905.035455] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.035777] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.036344] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 905.036804] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d51e44a-fba0-4322-a745-80c92a4076a1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.043465] env[62692]: DEBUG oslo_vmware.api [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 905.043465] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52438ccd-3aeb-e297-44e6-768ee1f41269" [ 905.043465] env[62692]: _type = "Task" [ 905.043465] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.061979] env[62692]: DEBUG oslo_vmware.api [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52438ccd-3aeb-e297-44e6-768ee1f41269, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.103337] env[62692]: DEBUG oslo_concurrency.lockutils [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.613s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.104346] env[62692]: DEBUG nova.compute.manager [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 905.115168] env[62692]: DEBUG oslo_concurrency.lockutils [None req-fd1fe936-78a2-4264-83e7-fd8f51817c1e tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.212s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.115168] env[62692]: DEBUG oslo_concurrency.lockutils [None req-fd1fe936-78a2-4264-83e7-fd8f51817c1e tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.115168] env[62692]: DEBUG oslo_concurrency.lockutils [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.665s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.115168] env[62692]: INFO nova.compute.claims [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 905.136356] env[62692]: DEBUG oslo_vmware.api [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141294, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06397} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.137167] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 905.138216] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d73b69a9-3cb3-4e3b-9199-3381dab3cb9a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.162081] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19/3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 905.165981] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aac218f6-47e9-4ae4-bce2-149a8daca4cb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.189736] env[62692]: DEBUG oslo_vmware.api [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 905.189736] env[62692]: value = "task-1141295" [ 905.189736] env[62692]: _type = "Task" [ 905.189736] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.200035] env[62692]: DEBUG oslo_vmware.api [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141295, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.437413] env[62692]: DEBUG nova.network.neutron [req-b7157021-eda6-4072-a041-7693d1184e0c req-08a174bd-a288-4c6b-b8fb-905369203573 service nova] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Updated VIF entry in instance network info cache for port f6f97326-a148-4785-a942-d2869ddffca3. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 905.437413] env[62692]: DEBUG nova.network.neutron [req-b7157021-eda6-4072-a041-7693d1184e0c req-08a174bd-a288-4c6b-b8fb-905369203573 service nova] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Updating instance_info_cache with network_info: [{"id": "f6f97326-a148-4785-a942-d2869ddffca3", "address": "fa:16:3e:9e:b5:42", "network": {"id": "14d97ab8-1b79-4ef2-a354-6398d00bdeba", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-707295669-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fc06bb980c584d2980ee5eb48e678772", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6f97326-a1", "ovs_interfaceid": "f6f97326-a148-4785-a942-d2869ddffca3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.562698] env[62692]: DEBUG oslo_vmware.api [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52438ccd-3aeb-e297-44e6-768ee1f41269, 'name': SearchDatastore_Task, 'duration_secs': 0.060639} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.564207] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.564207] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 905.564207] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.564207] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.564207] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 905.564434] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-39922e50-6417-40fc-92f3-14949f96dfcf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.578678] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 905.578882] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 905.579814] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e7cdd0d-e563-4233-a1d4-91a0aec23873 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.588270] env[62692]: DEBUG oslo_vmware.api [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 905.588270] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52c3eccc-f57f-1477-d593-0f0c9740f0f0" [ 905.588270] env[62692]: _type = "Task" [ 905.588270] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.603216] env[62692]: DEBUG oslo_vmware.api [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52c3eccc-f57f-1477-d593-0f0c9740f0f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.610181] env[62692]: DEBUG nova.compute.utils [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 905.611675] env[62692]: DEBUG nova.compute.manager [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 905.611873] env[62692]: DEBUG nova.network.neutron [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 905.640208] env[62692]: DEBUG oslo_concurrency.lockutils [None req-fd1fe936-78a2-4264-83e7-fd8f51817c1e tempest-ServerActionsV293TestJSON-1739480486 tempest-ServerActionsV293TestJSON-1739480486-project-member] Lock "81ce340c-fbef-4932-983c-595843530dbc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.429s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.671498] env[62692]: DEBUG nova.policy [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a28f5e0e0af443c5be82214d446c221d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '589895ae20394e2fa525707bf7d09f2a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 905.702015] env[62692]: DEBUG oslo_vmware.api [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141295, 'name': ReconfigVM_Task, 'duration_secs': 0.411108} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.702015] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19/3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 905.702015] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-650db828-7f4b-4ea2-bf65-7fd82bac5e36 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.709683] env[62692]: DEBUG oslo_vmware.api [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 905.709683] env[62692]: value = "task-1141296" [ 905.709683] env[62692]: _type = "Task" [ 905.709683] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.720087] env[62692]: DEBUG oslo_vmware.api [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141296, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.941572] env[62692]: DEBUG oslo_concurrency.lockutils [req-b7157021-eda6-4072-a041-7693d1184e0c req-08a174bd-a288-4c6b-b8fb-905369203573 service nova] Releasing lock "refresh_cache-ae44f074-a5c8-4259-99c6-9ce290b1570b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.070741] env[62692]: DEBUG nova.network.neutron [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Successfully created port: 2d5b5313-b87f-4cb1-9478-6fbbae387430 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 906.100016] env[62692]: DEBUG oslo_vmware.api [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52c3eccc-f57f-1477-d593-0f0c9740f0f0, 'name': SearchDatastore_Task, 'duration_secs': 0.014736} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.101240] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d80aecc-59e3-430a-ae7a-a13c4cdc3ee3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.107501] env[62692]: DEBUG oslo_vmware.api [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 906.107501] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52bf78e8-210c-136b-b42e-fe82a4c015ee" [ 906.107501] env[62692]: _type = "Task" [ 906.107501] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.116770] env[62692]: DEBUG nova.compute.manager [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 906.120272] env[62692]: DEBUG oslo_vmware.api [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52bf78e8-210c-136b-b42e-fe82a4c015ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.223562] env[62692]: DEBUG oslo_vmware.api [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141296, 'name': Rename_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.607552] env[62692]: DEBUG nova.network.neutron [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Successfully updated port: bfd79ef9-4a10-444c-96eb-1a4f4273a5d7 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 906.624642] env[62692]: DEBUG oslo_vmware.api [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52bf78e8-210c-136b-b42e-fe82a4c015ee, 'name': SearchDatastore_Task, 'duration_secs': 0.009543} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.624642] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.624804] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] ae44f074-a5c8-4259-99c6-9ce290b1570b/ae44f074-a5c8-4259-99c6-9ce290b1570b.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 906.630712] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3df2f5e5-d4f8-46f1-b957-e9bbbb911203 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.648997] env[62692]: DEBUG oslo_vmware.api [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 906.648997] env[62692]: value = "task-1141297" [ 906.648997] env[62692]: _type = "Task" [ 906.648997] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.665290] env[62692]: DEBUG oslo_vmware.api [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1141297, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.678023] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab190344-7cf7-4fa4-b67c-ad7c3a0736ba {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.691998] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1ddd4e4-c962-4fb9-b7b7-59851b33e902 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.728695] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c15e54fd-6518-467c-837b-033604b5fa0a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.746412] env[62692]: DEBUG oslo_vmware.api [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141296, 'name': Rename_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.747754] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3940555-b7c2-4214-9a3b-bd96cb7c021a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.754350] env[62692]: DEBUG nova.compute.manager [req-ad615f78-1e99-4efb-a962-989072316a4c req-f9ca249d-8b72-43f5-8d05-8bf99f00b7ad service nova] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Received event network-vif-plugged-bfd79ef9-4a10-444c-96eb-1a4f4273a5d7 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 906.754566] env[62692]: DEBUG oslo_concurrency.lockutils [req-ad615f78-1e99-4efb-a962-989072316a4c req-f9ca249d-8b72-43f5-8d05-8bf99f00b7ad service nova] Acquiring lock "9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.754773] env[62692]: DEBUG oslo_concurrency.lockutils [req-ad615f78-1e99-4efb-a962-989072316a4c req-f9ca249d-8b72-43f5-8d05-8bf99f00b7ad service nova] Lock "9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.754946] env[62692]: DEBUG oslo_concurrency.lockutils [req-ad615f78-1e99-4efb-a962-989072316a4c req-f9ca249d-8b72-43f5-8d05-8bf99f00b7ad service nova] Lock "9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.755147] env[62692]: DEBUG nova.compute.manager [req-ad615f78-1e99-4efb-a962-989072316a4c req-f9ca249d-8b72-43f5-8d05-8bf99f00b7ad service nova] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] No waiting events found dispatching network-vif-plugged-bfd79ef9-4a10-444c-96eb-1a4f4273a5d7 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 906.755325] env[62692]: WARNING nova.compute.manager [req-ad615f78-1e99-4efb-a962-989072316a4c req-f9ca249d-8b72-43f5-8d05-8bf99f00b7ad service nova] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Received unexpected event network-vif-plugged-bfd79ef9-4a10-444c-96eb-1a4f4273a5d7 for instance with vm_state building and task_state spawning. [ 906.767848] env[62692]: DEBUG nova.compute.provider_tree [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 907.138183] env[62692]: DEBUG nova.compute.manager [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 907.165493] env[62692]: DEBUG oslo_vmware.api [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1141297, 'name': CopyVirtualDisk_Task} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.166141] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] ae44f074-a5c8-4259-99c6-9ce290b1570b/ae44f074-a5c8-4259-99c6-9ce290b1570b.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 907.166835] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 907.166835] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6efc87c7-1eda-4374-897d-5554575d508a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.190483] env[62692]: DEBUG nova.virt.hardware [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 907.190483] env[62692]: DEBUG nova.virt.hardware [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 907.190483] env[62692]: DEBUG nova.virt.hardware [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 907.190483] env[62692]: DEBUG nova.virt.hardware [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 907.191492] env[62692]: DEBUG nova.virt.hardware [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 907.192607] env[62692]: DEBUG nova.virt.hardware [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 907.195025] env[62692]: DEBUG nova.virt.hardware [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 907.195025] env[62692]: DEBUG nova.virt.hardware [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 907.195025] env[62692]: DEBUG nova.virt.hardware [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 907.195025] env[62692]: DEBUG nova.virt.hardware [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 907.195025] env[62692]: DEBUG nova.virt.hardware [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 907.196013] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-506148b4-a77a-4ed8-a7b3-35dd10a05df5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.203282] env[62692]: DEBUG oslo_vmware.api [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 907.203282] env[62692]: value = "task-1141298" [ 907.203282] env[62692]: _type = "Task" [ 907.203282] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.205853] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "46bce865-188b-487c-a73f-cf524059c797" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.206150] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "46bce865-188b-487c-a73f-cf524059c797" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.214428] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b64b0dde-0382-4af5-bab4-af6710f320c3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.222747] env[62692]: DEBUG oslo_vmware.api [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1141298, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.244321] env[62692]: DEBUG oslo_vmware.api [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141296, 'name': Rename_Task, 'duration_secs': 1.154175} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.244321] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 907.244321] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0dfc0e90-4e30-4274-8d20-7693ef2d0706 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.255893] env[62692]: DEBUG oslo_vmware.api [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 907.255893] env[62692]: value = "task-1141299" [ 907.255893] env[62692]: _type = "Task" [ 907.255893] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.267863] env[62692]: DEBUG oslo_vmware.api [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141299, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.297250] env[62692]: ERROR nova.scheduler.client.report [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [req-aaaaabc7-0f63-4f9b-b60e-f4edee73a31d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 470758a8-5efa-4875-aad5-f512a727752f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-aaaaabc7-0f63-4f9b-b60e-f4edee73a31d"}]} [ 907.317794] env[62692]: DEBUG nova.scheduler.client.report [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Refreshing inventories for resource provider 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 907.333093] env[62692]: DEBUG nova.scheduler.client.report [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Updating ProviderTree inventory for provider 470758a8-5efa-4875-aad5-f512a727752f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 907.333365] env[62692]: DEBUG nova.compute.provider_tree [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 907.348507] env[62692]: DEBUG nova.scheduler.client.report [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Refreshing aggregate associations for resource provider 470758a8-5efa-4875-aad5-f512a727752f, aggregates: None {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 907.366239] env[62692]: DEBUG nova.scheduler.client.report [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Refreshing trait associations for resource provider 470758a8-5efa-4875-aad5-f512a727752f, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 907.720445] env[62692]: DEBUG oslo_vmware.api [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1141298, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.11912} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.720744] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 907.721922] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9583ecc-af25-43c9-a105-e519f83bfccc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.749959] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] ae44f074-a5c8-4259-99c6-9ce290b1570b/ae44f074-a5c8-4259-99c6-9ce290b1570b.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 907.754012] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d0602ae-892d-44cb-a044-9d84a0a35a38 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.782297] env[62692]: DEBUG oslo_vmware.api [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141299, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.783981] env[62692]: DEBUG oslo_vmware.api [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 907.783981] env[62692]: value = "task-1141300" [ 907.783981] env[62692]: _type = "Task" [ 907.783981] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.796506] env[62692]: DEBUG oslo_vmware.api [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1141300, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.909470] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a81030f8-9f59-487e-b736-90c382357582 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.920022] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b08adfb-1c49-4dbd-b75f-8893b3279007 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.957545] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e47fd37-0b68-4882-8e28-6d41f205a529 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.967912] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bdd1a53-2d9f-4916-807c-f097c3199752 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.985419] env[62692]: DEBUG nova.compute.provider_tree [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 908.115798] env[62692]: DEBUG nova.compute.manager [req-1c568348-455e-437e-8561-a761ad45e847 req-aad43cb7-26a5-4811-bd47-3d0cf5d2755d service nova] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Received event network-vif-plugged-2d5b5313-b87f-4cb1-9478-6fbbae387430 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 908.116460] env[62692]: DEBUG oslo_concurrency.lockutils [req-1c568348-455e-437e-8561-a761ad45e847 req-aad43cb7-26a5-4811-bd47-3d0cf5d2755d service nova] Acquiring lock "3fd2959b-80cb-470c-8c2a-40b7630458ed-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.116460] env[62692]: DEBUG oslo_concurrency.lockutils [req-1c568348-455e-437e-8561-a761ad45e847 req-aad43cb7-26a5-4811-bd47-3d0cf5d2755d service nova] Lock "3fd2959b-80cb-470c-8c2a-40b7630458ed-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.116460] env[62692]: DEBUG oslo_concurrency.lockutils [req-1c568348-455e-437e-8561-a761ad45e847 req-aad43cb7-26a5-4811-bd47-3d0cf5d2755d service nova] Lock "3fd2959b-80cb-470c-8c2a-40b7630458ed-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.116666] env[62692]: DEBUG nova.compute.manager [req-1c568348-455e-437e-8561-a761ad45e847 req-aad43cb7-26a5-4811-bd47-3d0cf5d2755d service nova] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] No waiting events found dispatching network-vif-plugged-2d5b5313-b87f-4cb1-9478-6fbbae387430 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 908.116703] env[62692]: WARNING nova.compute.manager [req-1c568348-455e-437e-8561-a761ad45e847 req-aad43cb7-26a5-4811-bd47-3d0cf5d2755d service nova] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Received unexpected event network-vif-plugged-2d5b5313-b87f-4cb1-9478-6fbbae387430 for instance with vm_state building and task_state spawning. [ 908.190213] env[62692]: DEBUG nova.network.neutron [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Successfully updated port: 2d5b5313-b87f-4cb1-9478-6fbbae387430 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 908.286850] env[62692]: DEBUG oslo_vmware.api [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141299, 'name': PowerOnVM_Task, 'duration_secs': 0.772243} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.290184] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 908.291128] env[62692]: INFO nova.compute.manager [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Took 8.93 seconds to spawn the instance on the hypervisor. [ 908.291128] env[62692]: DEBUG nova.compute.manager [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 908.291471] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3def7902-97fc-4d79-ace2-98568bf27a6f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.299829] env[62692]: DEBUG oslo_vmware.api [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1141300, 'name': ReconfigVM_Task, 'duration_secs': 0.381584} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.301648] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Reconfigured VM instance instance-00000041 to attach disk [datastore2] ae44f074-a5c8-4259-99c6-9ce290b1570b/ae44f074-a5c8-4259-99c6-9ce290b1570b.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 908.305065] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9725d8d0-b124-4965-99d4-e9c8fb43bdd6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.314177] env[62692]: DEBUG oslo_vmware.api [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 908.314177] env[62692]: value = "task-1141301" [ 908.314177] env[62692]: _type = "Task" [ 908.314177] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.323346] env[62692]: DEBUG oslo_vmware.api [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1141301, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.391554] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5c4e580d-f4ee-4b4c-9b15-56a1b2ea13e1 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "754b3c65-1e4b-49d2-8980-095d975edb01" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.391554] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5c4e580d-f4ee-4b4c-9b15-56a1b2ea13e1 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "754b3c65-1e4b-49d2-8980-095d975edb01" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.518693] env[62692]: DEBUG nova.scheduler.client.report [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Updated inventory for provider 470758a8-5efa-4875-aad5-f512a727752f with generation 93 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 908.518978] env[62692]: DEBUG nova.compute.provider_tree [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Updating resource provider 470758a8-5efa-4875-aad5-f512a727752f generation from 93 to 94 during operation: update_inventory {{(pid=62692) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 908.519304] env[62692]: DEBUG nova.compute.provider_tree [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 908.694233] env[62692]: DEBUG oslo_concurrency.lockutils [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "refresh_cache-3fd2959b-80cb-470c-8c2a-40b7630458ed" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.694426] env[62692]: DEBUG oslo_concurrency.lockutils [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired lock "refresh_cache-3fd2959b-80cb-470c-8c2a-40b7630458ed" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.694587] env[62692]: DEBUG nova.network.neutron [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 908.825197] env[62692]: INFO nova.compute.manager [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Took 54.26 seconds to build instance. [ 908.833153] env[62692]: DEBUG oslo_vmware.api [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1141301, 'name': Rename_Task, 'duration_secs': 0.147055} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.833602] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 908.833990] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6c383155-1a5a-4da8-b9db-540382609846 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.847495] env[62692]: DEBUG oslo_vmware.api [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 908.847495] env[62692]: value = "task-1141302" [ 908.847495] env[62692]: _type = "Task" [ 908.847495] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.858994] env[62692]: DEBUG oslo_vmware.api [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1141302, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.896751] env[62692]: DEBUG nova.compute.utils [None req-5c4e580d-f4ee-4b4c-9b15-56a1b2ea13e1 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 909.029059] env[62692]: DEBUG oslo_concurrency.lockutils [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.917s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.029059] env[62692]: DEBUG nova.compute.manager [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 909.032198] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.782s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.034149] env[62692]: INFO nova.compute.claims [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 909.150172] env[62692]: DEBUG nova.compute.manager [req-8b2b92ed-05a9-424a-8d31-3e278693397d req-215da0dd-340b-4dd4-9b54-d38c2ede6429 service nova] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Received event network-changed-bfd79ef9-4a10-444c-96eb-1a4f4273a5d7 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 909.150541] env[62692]: DEBUG nova.compute.manager [req-8b2b92ed-05a9-424a-8d31-3e278693397d req-215da0dd-340b-4dd4-9b54-d38c2ede6429 service nova] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Refreshing instance network info cache due to event network-changed-bfd79ef9-4a10-444c-96eb-1a4f4273a5d7. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 909.150883] env[62692]: DEBUG oslo_concurrency.lockutils [req-8b2b92ed-05a9-424a-8d31-3e278693397d req-215da0dd-340b-4dd4-9b54-d38c2ede6429 service nova] Acquiring lock "refresh_cache-9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.152734] env[62692]: DEBUG oslo_concurrency.lockutils [req-8b2b92ed-05a9-424a-8d31-3e278693397d req-215da0dd-340b-4dd4-9b54-d38c2ede6429 service nova] Acquired lock "refresh_cache-9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.152734] env[62692]: DEBUG nova.network.neutron [req-8b2b92ed-05a9-424a-8d31-3e278693397d req-215da0dd-340b-4dd4-9b54-d38c2ede6429 service nova] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Refreshing network info cache for port bfd79ef9-4a10-444c-96eb-1a4f4273a5d7 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 909.233165] env[62692]: DEBUG nova.network.neutron [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 909.328051] env[62692]: DEBUG oslo_concurrency.lockutils [None req-524b263a-5c79-45d1-b756-0f2c18b7d627 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.716s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.362099] env[62692]: DEBUG oslo_vmware.api [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1141302, 'name': PowerOnVM_Task, 'duration_secs': 0.446931} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.362389] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 909.362594] env[62692]: INFO nova.compute.manager [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Took 7.44 seconds to spawn the instance on the hypervisor. [ 909.362784] env[62692]: DEBUG nova.compute.manager [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 909.363915] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fa2bc06-fe4a-48e2-be7f-1ec3d52c6e8b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.399028] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5c4e580d-f4ee-4b4c-9b15-56a1b2ea13e1 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "754b3c65-1e4b-49d2-8980-095d975edb01" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.407462] env[62692]: DEBUG nova.network.neutron [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Updating instance_info_cache with network_info: [{"id": "2d5b5313-b87f-4cb1-9478-6fbbae387430", "address": "fa:16:3e:9c:ef:2e", "network": {"id": "dd260020-2d74-46ba-aa78-af29bdefa176", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-617179655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "589895ae20394e2fa525707bf7d09f2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d5b5313-b8", "ovs_interfaceid": "2d5b5313-b87f-4cb1-9478-6fbbae387430", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.542367] env[62692]: DEBUG nova.compute.utils [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 909.544021] env[62692]: DEBUG nova.compute.manager [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 909.546639] env[62692]: DEBUG nova.network.neutron [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 909.608193] env[62692]: DEBUG nova.network.neutron [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Successfully updated port: b81f3e56-2a23-40b7-8153-e53be84131ca {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 909.677351] env[62692]: DEBUG nova.policy [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '249a3b566fcf4d88a973693512df370e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0b0b2e09a06a4fc7a00a03f72989ffed', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 909.699860] env[62692]: DEBUG nova.network.neutron [req-8b2b92ed-05a9-424a-8d31-3e278693397d req-215da0dd-340b-4dd4-9b54-d38c2ede6429 service nova] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 909.838056] env[62692]: DEBUG nova.compute.manager [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 909.855612] env[62692]: DEBUG nova.network.neutron [req-8b2b92ed-05a9-424a-8d31-3e278693397d req-215da0dd-340b-4dd4-9b54-d38c2ede6429 service nova] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.886413] env[62692]: INFO nova.compute.manager [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Took 54.01 seconds to build instance. [ 909.913021] env[62692]: DEBUG oslo_concurrency.lockutils [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Releasing lock "refresh_cache-3fd2959b-80cb-470c-8c2a-40b7630458ed" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.913021] env[62692]: DEBUG nova.compute.manager [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Instance network_info: |[{"id": "2d5b5313-b87f-4cb1-9478-6fbbae387430", "address": "fa:16:3e:9c:ef:2e", "network": {"id": "dd260020-2d74-46ba-aa78-af29bdefa176", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-617179655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "589895ae20394e2fa525707bf7d09f2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d5b5313-b8", "ovs_interfaceid": "2d5b5313-b87f-4cb1-9478-6fbbae387430", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 909.913021] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9c:ef:2e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55764410-260e-4339-a020-6b30995584bf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2d5b5313-b87f-4cb1-9478-6fbbae387430', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 909.923362] env[62692]: DEBUG oslo.service.loopingcall [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 909.924725] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 909.926245] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9ae693cd-4d1f-4541-b25f-b6aad9661d85 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.952276] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 909.952276] env[62692]: value = "task-1141303" [ 909.952276] env[62692]: _type = "Task" [ 909.952276] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.966105] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141303, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.045688] env[62692]: DEBUG nova.compute.manager [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 910.109779] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Acquiring lock "refresh_cache-9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.246555] env[62692]: DEBUG nova.network.neutron [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Successfully created port: 6702eb8d-7d7d-4384-8eb4-d2196df920ef {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 910.360580] env[62692]: DEBUG oslo_concurrency.lockutils [req-8b2b92ed-05a9-424a-8d31-3e278693397d req-215da0dd-340b-4dd4-9b54-d38c2ede6429 service nova] Releasing lock "refresh_cache-9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.361407] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Acquired lock "refresh_cache-9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.361576] env[62692]: DEBUG nova.network.neutron [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 910.363987] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.391268] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9d634af6-6164-4199-8a6e-bbb31e844d2c tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lock "ae44f074-a5c8-4259-99c6-9ce290b1570b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.489s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.463700] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141303, 'name': CreateVM_Task, 'duration_secs': 0.425774} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.466373] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 910.467686] env[62692]: DEBUG oslo_concurrency.lockutils [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.467686] env[62692]: DEBUG oslo_concurrency.lockutils [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.470096] env[62692]: DEBUG oslo_concurrency.lockutils [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 910.470096] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-844deb7b-96b9-4157-bcc9-036db3373723 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.476030] env[62692]: DEBUG oslo_vmware.api [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 910.476030] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]525cd08a-7177-73c2-0a65-7cf58be52fbf" [ 910.476030] env[62692]: _type = "Task" [ 910.476030] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.486607] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5c4e580d-f4ee-4b4c-9b15-56a1b2ea13e1 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "754b3c65-1e4b-49d2-8980-095d975edb01" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.486837] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5c4e580d-f4ee-4b4c-9b15-56a1b2ea13e1 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "754b3c65-1e4b-49d2-8980-095d975edb01" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.488792] env[62692]: INFO nova.compute.manager [None req-5c4e580d-f4ee-4b4c-9b15-56a1b2ea13e1 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Attaching volume 57d15e0c-bb59-47d1-abf1-252dc8cb294a to /dev/sdb [ 910.505529] env[62692]: DEBUG oslo_vmware.api [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]525cd08a-7177-73c2-0a65-7cf58be52fbf, 'name': SearchDatastore_Task, 'duration_secs': 0.011909} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.508586] env[62692]: DEBUG oslo_concurrency.lockutils [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.508888] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 910.509284] env[62692]: DEBUG oslo_concurrency.lockutils [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.509912] env[62692]: DEBUG oslo_concurrency.lockutils [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.509912] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 910.510436] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2e7c6e03-68c1-432b-ba8d-e7647f4f0b39 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.520551] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 910.520745] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 910.523949] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88fa8214-22e9-4c23-a9ce-5ee0ae76317f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.528427] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ab1f37d-6c6a-4fc5-8765-8f8c287cc193 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.534789] env[62692]: DEBUG oslo_vmware.api [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 910.534789] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52b66658-a124-a23f-5c78-f53e3407ec5e" [ 910.534789] env[62692]: _type = "Task" [ 910.534789] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.540669] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d22b73-7b8e-4d8d-80a2-ff45f5808c75 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.555145] env[62692]: DEBUG oslo_vmware.api [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52b66658-a124-a23f-5c78-f53e3407ec5e, 'name': SearchDatastore_Task, 'duration_secs': 0.011386} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.562545] env[62692]: DEBUG nova.virt.block_device [None req-5c4e580d-f4ee-4b4c-9b15-56a1b2ea13e1 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Updating existing volume attachment record: 22e754ac-d291-4837-818e-64a6eae1a665 {{(pid=62692) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 910.567539] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49105490-1116-485a-8110-b8ff72f6e529 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.574810] env[62692]: DEBUG oslo_vmware.api [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 910.574810] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52ffe841-6436-c6d4-019b-0eb626608cfa" [ 910.574810] env[62692]: _type = "Task" [ 910.574810] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.585550] env[62692]: DEBUG oslo_vmware.api [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52ffe841-6436-c6d4-019b-0eb626608cfa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.638667] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0639afc-78d7-4e4a-b438-bffe947bfc5e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.655342] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-014ee8ef-c059-4835-850d-fa91080ca3c5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.690118] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4c01526-56bc-4395-99af-ec48780324c0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.698763] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41dd2607-d79d-4834-87a0-b611e25bab73 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.715957] env[62692]: DEBUG nova.compute.provider_tree [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 910.895429] env[62692]: DEBUG nova.compute.manager [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 910.904261] env[62692]: DEBUG nova.network.neutron [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 910.924815] env[62692]: DEBUG nova.compute.manager [req-1f39e875-c327-415f-8d39-6ab7570ca090 req-0f07f729-cc61-4641-94f9-8aa9194fe5be service nova] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Received event network-changed-2d5b5313-b87f-4cb1-9478-6fbbae387430 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 910.925054] env[62692]: DEBUG nova.compute.manager [req-1f39e875-c327-415f-8d39-6ab7570ca090 req-0f07f729-cc61-4641-94f9-8aa9194fe5be service nova] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Refreshing instance network info cache due to event network-changed-2d5b5313-b87f-4cb1-9478-6fbbae387430. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 910.925363] env[62692]: DEBUG oslo_concurrency.lockutils [req-1f39e875-c327-415f-8d39-6ab7570ca090 req-0f07f729-cc61-4641-94f9-8aa9194fe5be service nova] Acquiring lock "refresh_cache-3fd2959b-80cb-470c-8c2a-40b7630458ed" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.925497] env[62692]: DEBUG oslo_concurrency.lockutils [req-1f39e875-c327-415f-8d39-6ab7570ca090 req-0f07f729-cc61-4641-94f9-8aa9194fe5be service nova] Acquired lock "refresh_cache-3fd2959b-80cb-470c-8c2a-40b7630458ed" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.925679] env[62692]: DEBUG nova.network.neutron [req-1f39e875-c327-415f-8d39-6ab7570ca090 req-0f07f729-cc61-4641-94f9-8aa9194fe5be service nova] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Refreshing network info cache for port 2d5b5313-b87f-4cb1-9478-6fbbae387430 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 911.069086] env[62692]: DEBUG nova.compute.manager [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 911.095706] env[62692]: DEBUG oslo_vmware.api [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52ffe841-6436-c6d4-019b-0eb626608cfa, 'name': SearchDatastore_Task, 'duration_secs': 0.012199} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.097140] env[62692]: DEBUG oslo_concurrency.lockutils [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.097521] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 3fd2959b-80cb-470c-8c2a-40b7630458ed/3fd2959b-80cb-470c-8c2a-40b7630458ed.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 911.097852] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3ef88eaf-9ccf-4ab5-903d-2b01091770c9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.109757] env[62692]: DEBUG oslo_vmware.api [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 911.109757] env[62692]: value = "task-1141307" [ 911.109757] env[62692]: _type = "Task" [ 911.109757] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.117312] env[62692]: DEBUG nova.virt.hardware [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='c204766882baa94d278d4f8e38543113',container_format='bare',created_at=2024-10-05T20:34:00Z,direct_url=,disk_format='vmdk',id=427c2d72-be26-42e4-9f8c-ca628806da41,min_disk=1,min_ram=0,name='tempest-test-snap-1999378371',owner='0b0b2e09a06a4fc7a00a03f72989ffed',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-05T20:34:16Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 911.117897] env[62692]: DEBUG nova.virt.hardware [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 911.118129] env[62692]: DEBUG nova.virt.hardware [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 911.118308] env[62692]: DEBUG nova.virt.hardware [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 911.118476] env[62692]: DEBUG nova.virt.hardware [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 911.118819] env[62692]: DEBUG nova.virt.hardware [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 911.119118] env[62692]: DEBUG nova.virt.hardware [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 911.119335] env[62692]: DEBUG nova.virt.hardware [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 911.119759] env[62692]: DEBUG nova.virt.hardware [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 911.119933] env[62692]: DEBUG nova.virt.hardware [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 911.120219] env[62692]: DEBUG nova.virt.hardware [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 911.121400] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-525e768e-88c0-4fe9-a6b8-188de3bfd1d6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.136077] env[62692]: DEBUG oslo_vmware.api [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141307, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.136981] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07d1d571-4811-4a9f-8b77-18006375b94b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.221779] env[62692]: DEBUG nova.scheduler.client.report [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 911.316137] env[62692]: DEBUG nova.network.neutron [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Updating instance_info_cache with network_info: [{"id": "bfd79ef9-4a10-444c-96eb-1a4f4273a5d7", "address": "fa:16:3e:83:2a:f6", "network": {"id": "4c74b409-989a-4073-9def-fc9e2e3cc154", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1736671348", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.146", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fed583a514c4043a4b5772c73e21aff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31e77685-b4dd-4810-80ef-24115ea9ea62", "external-id": "nsx-vlan-transportzone-56", "segmentation_id": 56, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfd79ef9-4a", "ovs_interfaceid": "bfd79ef9-4a10-444c-96eb-1a4f4273a5d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b81f3e56-2a23-40b7-8153-e53be84131ca", "address": "fa:16:3e:a3:88:06", "network": {"id": "e92d05d2-f951-4463-ac95-62df576da43e", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1836680878", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.40", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4fed583a514c4043a4b5772c73e21aff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd3c6b64-aba2-4bdc-a693-3b4dff3ed861", "external-id": "nsx-vlan-transportzone-600", "segmentation_id": 600, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb81f3e56-2a", "ovs_interfaceid": "b81f3e56-2a23-40b7-8153-e53be84131ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.413430] env[62692]: DEBUG nova.compute.manager [req-f3935fb3-3fc0-4382-8d72-2e9ef8845cb2 req-f4bcba8a-d0ef-456a-a936-0ac2b5fd693c service nova] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Received event network-vif-plugged-b81f3e56-2a23-40b7-8153-e53be84131ca {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 911.413430] env[62692]: DEBUG oslo_concurrency.lockutils [req-f3935fb3-3fc0-4382-8d72-2e9ef8845cb2 req-f4bcba8a-d0ef-456a-a936-0ac2b5fd693c service nova] Acquiring lock "9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.413430] env[62692]: DEBUG oslo_concurrency.lockutils [req-f3935fb3-3fc0-4382-8d72-2e9ef8845cb2 req-f4bcba8a-d0ef-456a-a936-0ac2b5fd693c service nova] Lock "9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.413567] env[62692]: DEBUG oslo_concurrency.lockutils [req-f3935fb3-3fc0-4382-8d72-2e9ef8845cb2 req-f4bcba8a-d0ef-456a-a936-0ac2b5fd693c service nova] Lock "9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.413739] env[62692]: DEBUG nova.compute.manager [req-f3935fb3-3fc0-4382-8d72-2e9ef8845cb2 req-f4bcba8a-d0ef-456a-a936-0ac2b5fd693c service nova] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] No waiting events found dispatching network-vif-plugged-b81f3e56-2a23-40b7-8153-e53be84131ca {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 911.413939] env[62692]: WARNING nova.compute.manager [req-f3935fb3-3fc0-4382-8d72-2e9ef8845cb2 req-f4bcba8a-d0ef-456a-a936-0ac2b5fd693c service nova] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Received unexpected event network-vif-plugged-b81f3e56-2a23-40b7-8153-e53be84131ca for instance with vm_state building and task_state spawning. [ 911.414158] env[62692]: DEBUG nova.compute.manager [req-f3935fb3-3fc0-4382-8d72-2e9ef8845cb2 req-f4bcba8a-d0ef-456a-a936-0ac2b5fd693c service nova] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Received event network-changed-b81f3e56-2a23-40b7-8153-e53be84131ca {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 911.414878] env[62692]: DEBUG nova.compute.manager [req-f3935fb3-3fc0-4382-8d72-2e9ef8845cb2 req-f4bcba8a-d0ef-456a-a936-0ac2b5fd693c service nova] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Refreshing instance network info cache due to event network-changed-b81f3e56-2a23-40b7-8153-e53be84131ca. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 911.414878] env[62692]: DEBUG oslo_concurrency.lockutils [req-f3935fb3-3fc0-4382-8d72-2e9ef8845cb2 req-f4bcba8a-d0ef-456a-a936-0ac2b5fd693c service nova] Acquiring lock "refresh_cache-9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.416970] env[62692]: DEBUG oslo_concurrency.lockutils [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.622184] env[62692]: DEBUG oslo_vmware.api [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141307, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.726770] env[62692]: DEBUG nova.network.neutron [req-1f39e875-c327-415f-8d39-6ab7570ca090 req-0f07f729-cc61-4641-94f9-8aa9194fe5be service nova] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Updated VIF entry in instance network info cache for port 2d5b5313-b87f-4cb1-9478-6fbbae387430. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 911.727372] env[62692]: DEBUG nova.network.neutron [req-1f39e875-c327-415f-8d39-6ab7570ca090 req-0f07f729-cc61-4641-94f9-8aa9194fe5be service nova] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Updating instance_info_cache with network_info: [{"id": "2d5b5313-b87f-4cb1-9478-6fbbae387430", "address": "fa:16:3e:9c:ef:2e", "network": {"id": "dd260020-2d74-46ba-aa78-af29bdefa176", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-617179655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "589895ae20394e2fa525707bf7d09f2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d5b5313-b8", "ovs_interfaceid": "2d5b5313-b87f-4cb1-9478-6fbbae387430", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.730189] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.698s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.734342] env[62692]: DEBUG nova.compute.manager [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 911.737318] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.270s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.737318] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.739994] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.036s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.745056] env[62692]: INFO nova.compute.claims [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 911.790279] env[62692]: INFO nova.scheduler.client.report [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Deleted allocations for instance 6b2d3385-a9af-432b-a1e7-c91adb2e1f69 [ 911.818710] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Releasing lock "refresh_cache-9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.819648] env[62692]: DEBUG nova.compute.manager [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Instance network_info: |[{"id": "bfd79ef9-4a10-444c-96eb-1a4f4273a5d7", "address": "fa:16:3e:83:2a:f6", "network": {"id": "4c74b409-989a-4073-9def-fc9e2e3cc154", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1736671348", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.146", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fed583a514c4043a4b5772c73e21aff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31e77685-b4dd-4810-80ef-24115ea9ea62", "external-id": "nsx-vlan-transportzone-56", "segmentation_id": 56, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfd79ef9-4a", "ovs_interfaceid": "bfd79ef9-4a10-444c-96eb-1a4f4273a5d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b81f3e56-2a23-40b7-8153-e53be84131ca", "address": "fa:16:3e:a3:88:06", "network": {"id": "e92d05d2-f951-4463-ac95-62df576da43e", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1836680878", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.40", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4fed583a514c4043a4b5772c73e21aff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd3c6b64-aba2-4bdc-a693-3b4dff3ed861", "external-id": "nsx-vlan-transportzone-600", "segmentation_id": 600, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb81f3e56-2a", "ovs_interfaceid": "b81f3e56-2a23-40b7-8153-e53be84131ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 911.819648] env[62692]: DEBUG oslo_concurrency.lockutils [req-f3935fb3-3fc0-4382-8d72-2e9ef8845cb2 req-f4bcba8a-d0ef-456a-a936-0ac2b5fd693c service nova] Acquired lock "refresh_cache-9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.819953] env[62692]: DEBUG nova.network.neutron [req-f3935fb3-3fc0-4382-8d72-2e9ef8845cb2 req-f4bcba8a-d0ef-456a-a936-0ac2b5fd693c service nova] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Refreshing network info cache for port b81f3e56-2a23-40b7-8153-e53be84131ca {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 911.824539] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:83:2a:f6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '31e77685-b4dd-4810-80ef-24115ea9ea62', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bfd79ef9-4a10-444c-96eb-1a4f4273a5d7', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:a3:88:06', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bd3c6b64-aba2-4bdc-a693-3b4dff3ed861', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b81f3e56-2a23-40b7-8153-e53be84131ca', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 911.835421] env[62692]: DEBUG oslo.service.loopingcall [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 911.836629] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 911.836887] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-374b1dfa-69cf-4307-aa0a-a5c2ef0e8a98 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.864247] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 911.864247] env[62692]: value = "task-1141308" [ 911.864247] env[62692]: _type = "Task" [ 911.864247] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.879125] env[62692]: DEBUG oslo_concurrency.lockutils [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Acquiring lock "084ce29c-302a-45cc-b36f-0143eccee298" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.879501] env[62692]: DEBUG oslo_concurrency.lockutils [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Lock "084ce29c-302a-45cc-b36f-0143eccee298" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.879978] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141308, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.067141] env[62692]: DEBUG nova.network.neutron [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Successfully updated port: 6702eb8d-7d7d-4384-8eb4-d2196df920ef {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 912.124776] env[62692]: DEBUG oslo_vmware.api [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141307, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.238098] env[62692]: DEBUG oslo_concurrency.lockutils [req-1f39e875-c327-415f-8d39-6ab7570ca090 req-0f07f729-cc61-4641-94f9-8aa9194fe5be service nova] Releasing lock "refresh_cache-3fd2959b-80cb-470c-8c2a-40b7630458ed" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.252612] env[62692]: DEBUG nova.compute.utils [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 912.254405] env[62692]: DEBUG nova.compute.manager [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Not allocating networking since 'none' was specified. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 912.305650] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c345e06f-0c5a-496f-9c39-b7668e079d07 tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Lock "6b2d3385-a9af-432b-a1e7-c91adb2e1f69" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.774s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.378317] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141308, 'name': CreateVM_Task, 'duration_secs': 0.478487} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.378614] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 912.380197] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.380811] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.381330] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 912.384628] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-145d612e-bf9e-466b-bc2b-8df6cf5d0e27 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.391941] env[62692]: DEBUG oslo_vmware.api [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Waiting for the task: (returnval){ [ 912.391941] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52772c93-0a1d-309b-b779-39010fe6c111" [ 912.391941] env[62692]: _type = "Task" [ 912.391941] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.401158] env[62692]: DEBUG oslo_vmware.api [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52772c93-0a1d-309b-b779-39010fe6c111, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.570151] env[62692]: DEBUG oslo_concurrency.lockutils [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "refresh_cache-44699df6-5262-4700-a096-8c0e450323d2" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.570358] env[62692]: DEBUG oslo_concurrency.lockutils [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquired lock "refresh_cache-44699df6-5262-4700-a096-8c0e450323d2" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.570519] env[62692]: DEBUG nova.network.neutron [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 912.624717] env[62692]: DEBUG oslo_vmware.api [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141307, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.310912} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.625022] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 3fd2959b-80cb-470c-8c2a-40b7630458ed/3fd2959b-80cb-470c-8c2a-40b7630458ed.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 912.625263] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 912.626546] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-09a34c3e-eb05-4c0e-8a68-ce0c83ee06dd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.629208] env[62692]: DEBUG nova.network.neutron [req-f3935fb3-3fc0-4382-8d72-2e9ef8845cb2 req-f4bcba8a-d0ef-456a-a936-0ac2b5fd693c service nova] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Updated VIF entry in instance network info cache for port b81f3e56-2a23-40b7-8153-e53be84131ca. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 912.629610] env[62692]: DEBUG nova.network.neutron [req-f3935fb3-3fc0-4382-8d72-2e9ef8845cb2 req-f4bcba8a-d0ef-456a-a936-0ac2b5fd693c service nova] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Updating instance_info_cache with network_info: [{"id": "bfd79ef9-4a10-444c-96eb-1a4f4273a5d7", "address": "fa:16:3e:83:2a:f6", "network": {"id": "4c74b409-989a-4073-9def-fc9e2e3cc154", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1736671348", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.146", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fed583a514c4043a4b5772c73e21aff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31e77685-b4dd-4810-80ef-24115ea9ea62", "external-id": "nsx-vlan-transportzone-56", "segmentation_id": 56, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfd79ef9-4a", "ovs_interfaceid": "bfd79ef9-4a10-444c-96eb-1a4f4273a5d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b81f3e56-2a23-40b7-8153-e53be84131ca", "address": "fa:16:3e:a3:88:06", "network": {"id": "e92d05d2-f951-4463-ac95-62df576da43e", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1836680878", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.40", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4fed583a514c4043a4b5772c73e21aff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd3c6b64-aba2-4bdc-a693-3b4dff3ed861", "external-id": "nsx-vlan-transportzone-600", "segmentation_id": 600, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb81f3e56-2a", "ovs_interfaceid": "b81f3e56-2a23-40b7-8153-e53be84131ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.636810] env[62692]: DEBUG oslo_vmware.api [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 912.636810] env[62692]: value = "task-1141309" [ 912.636810] env[62692]: _type = "Task" [ 912.636810] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.660066] env[62692]: DEBUG oslo_vmware.api [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141309, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.758104] env[62692]: DEBUG nova.compute.manager [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 912.905684] env[62692]: DEBUG oslo_vmware.api [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52772c93-0a1d-309b-b779-39010fe6c111, 'name': SearchDatastore_Task, 'duration_secs': 0.012012} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.909927] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.910260] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 912.910549] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.910708] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.910893] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 912.912568] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e077a49f-7796-4f20-be0f-835dfdd8c84b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.928657] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 912.928874] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 912.929805] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99e52147-2402-461b-b213-7b04bdc62421 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.937333] env[62692]: DEBUG oslo_vmware.api [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Waiting for the task: (returnval){ [ 912.937333] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5225a76c-6f62-f80b-3f01-31bc8f326501" [ 912.937333] env[62692]: _type = "Task" [ 912.937333] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.952122] env[62692]: DEBUG oslo_vmware.api [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5225a76c-6f62-f80b-3f01-31bc8f326501, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.117719] env[62692]: DEBUG nova.network.neutron [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 913.128883] env[62692]: DEBUG nova.compute.manager [req-5d2e74de-296d-4f78-b225-ccb79f971051 req-288b3751-914a-4ca4-aa0e-9c78855d2d01 service nova] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Received event network-changed-2c7422d2-dfef-4010-a99a-60c510c54a3c {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 913.129177] env[62692]: DEBUG nova.compute.manager [req-5d2e74de-296d-4f78-b225-ccb79f971051 req-288b3751-914a-4ca4-aa0e-9c78855d2d01 service nova] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Refreshing instance network info cache due to event network-changed-2c7422d2-dfef-4010-a99a-60c510c54a3c. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 913.129418] env[62692]: DEBUG oslo_concurrency.lockutils [req-5d2e74de-296d-4f78-b225-ccb79f971051 req-288b3751-914a-4ca4-aa0e-9c78855d2d01 service nova] Acquiring lock "refresh_cache-3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.129570] env[62692]: DEBUG oslo_concurrency.lockutils [req-5d2e74de-296d-4f78-b225-ccb79f971051 req-288b3751-914a-4ca4-aa0e-9c78855d2d01 service nova] Acquired lock "refresh_cache-3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.129748] env[62692]: DEBUG nova.network.neutron [req-5d2e74de-296d-4f78-b225-ccb79f971051 req-288b3751-914a-4ca4-aa0e-9c78855d2d01 service nova] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Refreshing network info cache for port 2c7422d2-dfef-4010-a99a-60c510c54a3c {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 913.135915] env[62692]: DEBUG oslo_concurrency.lockutils [req-f3935fb3-3fc0-4382-8d72-2e9ef8845cb2 req-f4bcba8a-d0ef-456a-a936-0ac2b5fd693c service nova] Releasing lock "refresh_cache-9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.151533] env[62692]: DEBUG oslo_vmware.api [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141309, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.109302} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.152061] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 913.152865] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7cd91d9-07f0-4c0f-9eac-841c63d348fd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.178565] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 3fd2959b-80cb-470c-8c2a-40b7630458ed/3fd2959b-80cb-470c-8c2a-40b7630458ed.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 913.183253] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b7f52447-7729-4320-b634-54366d6b5ed9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.207918] env[62692]: DEBUG oslo_vmware.api [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 913.207918] env[62692]: value = "task-1141311" [ 913.207918] env[62692]: _type = "Task" [ 913.207918] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.219860] env[62692]: DEBUG oslo_vmware.api [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141311, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.281144] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa16609a-dbdb-46cf-8e78-14377c61b194 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.289900] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20cfec7a-9daf-4ad5-9b77-316016d88a45 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.326076] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cca28760-3fb0-4f8a-95dc-f093e099db15 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.336729] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85a87ad2-2262-4b32-ad39-0e6ca5550437 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.350519] env[62692]: DEBUG nova.compute.provider_tree [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 913.370329] env[62692]: DEBUG nova.network.neutron [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Updating instance_info_cache with network_info: [{"id": "6702eb8d-7d7d-4384-8eb4-d2196df920ef", "address": "fa:16:3e:ba:19:e9", "network": {"id": "2c82845f-101b-48c4-bd0b-4704573a3615", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1972972124-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b0b2e09a06a4fc7a00a03f72989ffed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16f09e8c-5240-4839-80cc-62ec29700bd2", "external-id": "nsx-vlan-transportzone-720", "segmentation_id": 720, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6702eb8d-7d", "ovs_interfaceid": "6702eb8d-7d7d-4384-8eb4-d2196df920ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.410777] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Acquiring lock "ae44f074-a5c8-4259-99c6-9ce290b1570b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.411060] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lock "ae44f074-a5c8-4259-99c6-9ce290b1570b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.411731] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Acquiring lock "ae44f074-a5c8-4259-99c6-9ce290b1570b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.411731] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lock "ae44f074-a5c8-4259-99c6-9ce290b1570b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.411731] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lock "ae44f074-a5c8-4259-99c6-9ce290b1570b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.414769] env[62692]: INFO nova.compute.manager [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Terminating instance [ 913.416992] env[62692]: DEBUG nova.compute.manager [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 913.417255] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 913.418166] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-482f2578-2d8b-402f-b414-0acb6f965864 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.426711] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 913.427083] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e095bfe5-a45f-40b2-8234-c48d12001dd3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.434792] env[62692]: DEBUG oslo_vmware.api [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 913.434792] env[62692]: value = "task-1141312" [ 913.434792] env[62692]: _type = "Task" [ 913.434792] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.456287] env[62692]: DEBUG oslo_vmware.api [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5225a76c-6f62-f80b-3f01-31bc8f326501, 'name': SearchDatastore_Task, 'duration_secs': 0.015899} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.456961] env[62692]: DEBUG oslo_vmware.api [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1141312, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.457838] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d25246c0-7692-45e4-aca0-2a9a2af6db3d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.464111] env[62692]: DEBUG oslo_vmware.api [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Waiting for the task: (returnval){ [ 913.464111] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52ac33a5-dafd-78b7-1a55-cfcc6671d7e5" [ 913.464111] env[62692]: _type = "Task" [ 913.464111] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.472976] env[62692]: DEBUG oslo_vmware.api [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52ac33a5-dafd-78b7-1a55-cfcc6671d7e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.717853] env[62692]: DEBUG oslo_vmware.api [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141311, 'name': ReconfigVM_Task, 'duration_secs': 0.331992} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.718202] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 3fd2959b-80cb-470c-8c2a-40b7630458ed/3fd2959b-80cb-470c-8c2a-40b7630458ed.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 913.718845] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-18a8b6fd-62fb-4c55-8d29-0463ecbc72e8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.726144] env[62692]: DEBUG oslo_vmware.api [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 913.726144] env[62692]: value = "task-1141313" [ 913.726144] env[62692]: _type = "Task" [ 913.726144] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.737033] env[62692]: DEBUG oslo_vmware.api [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141313, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.769635] env[62692]: DEBUG nova.compute.manager [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 913.799191] env[62692]: DEBUG nova.virt.hardware [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 913.799494] env[62692]: DEBUG nova.virt.hardware [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 913.799658] env[62692]: DEBUG nova.virt.hardware [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 913.799874] env[62692]: DEBUG nova.virt.hardware [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 913.800078] env[62692]: DEBUG nova.virt.hardware [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 913.800263] env[62692]: DEBUG nova.virt.hardware [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 913.800489] env[62692]: DEBUG nova.virt.hardware [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 913.800654] env[62692]: DEBUG nova.virt.hardware [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 913.801176] env[62692]: DEBUG nova.virt.hardware [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 913.801176] env[62692]: DEBUG nova.virt.hardware [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 913.801176] env[62692]: DEBUG nova.virt.hardware [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 913.802514] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90a0f86b-4120-4c51-aafc-b1572a586aad {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.811638] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96d436f2-766b-4d31-bd57-655a59239912 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.828225] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Instance VIF info [] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 913.833870] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Creating folder: Project (8aa905dd13704c5787c365b9db641999). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 913.834254] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-82dcf9cf-1575-4ffe-8208-9545ec1767e0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.846780] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Created folder: Project (8aa905dd13704c5787c365b9db641999) in parent group-v248868. [ 913.847101] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Creating folder: Instances. Parent ref: group-v248976. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 913.847367] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-30afa880-b8a6-4a01-bc1d-4e22d00b9320 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.853140] env[62692]: DEBUG nova.scheduler.client.report [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 913.861470] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Created folder: Instances in parent group-v248976. [ 913.861470] env[62692]: DEBUG oslo.service.loopingcall [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 913.861470] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 913.861470] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4e2ceec5-d6ce-4626-b3e2-fefada66ee66 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.873910] env[62692]: DEBUG oslo_concurrency.lockutils [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Releasing lock "refresh_cache-44699df6-5262-4700-a096-8c0e450323d2" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.874387] env[62692]: DEBUG nova.compute.manager [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Instance network_info: |[{"id": "6702eb8d-7d7d-4384-8eb4-d2196df920ef", "address": "fa:16:3e:ba:19:e9", "network": {"id": "2c82845f-101b-48c4-bd0b-4704573a3615", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1972972124-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b0b2e09a06a4fc7a00a03f72989ffed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16f09e8c-5240-4839-80cc-62ec29700bd2", "external-id": "nsx-vlan-transportzone-720", "segmentation_id": 720, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6702eb8d-7d", "ovs_interfaceid": "6702eb8d-7d7d-4384-8eb4-d2196df920ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 913.875184] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ba:19:e9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '16f09e8c-5240-4839-80cc-62ec29700bd2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6702eb8d-7d7d-4384-8eb4-d2196df920ef', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 913.883199] env[62692]: DEBUG oslo.service.loopingcall [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 913.884088] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 913.884334] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ea4f2247-9980-4c65-bbf1-59881c47213f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.901012] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 913.901012] env[62692]: value = "task-1141316" [ 913.901012] env[62692]: _type = "Task" [ 913.901012] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.906549] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 913.906549] env[62692]: value = "task-1141317" [ 913.906549] env[62692]: _type = "Task" [ 913.906549] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.914179] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141316, 'name': CreateVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.922022] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141317, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.948378] env[62692]: DEBUG oslo_vmware.api [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1141312, 'name': PowerOffVM_Task, 'duration_secs': 0.220842} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.948955] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 913.948955] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 913.949330] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b4f85498-af5d-4981-bb7d-2c1b1f5b063e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.975561] env[62692]: DEBUG oslo_vmware.api [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52ac33a5-dafd-78b7-1a55-cfcc6671d7e5, 'name': SearchDatastore_Task, 'duration_secs': 0.012348} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.976604] env[62692]: DEBUG nova.network.neutron [req-5d2e74de-296d-4f78-b225-ccb79f971051 req-288b3751-914a-4ca4-aa0e-9c78855d2d01 service nova] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Updated VIF entry in instance network info cache for port 2c7422d2-dfef-4010-a99a-60c510c54a3c. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 913.977329] env[62692]: DEBUG nova.network.neutron [req-5d2e74de-296d-4f78-b225-ccb79f971051 req-288b3751-914a-4ca4-aa0e-9c78855d2d01 service nova] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Updating instance_info_cache with network_info: [{"id": "2c7422d2-dfef-4010-a99a-60c510c54a3c", "address": "fa:16:3e:71:16:ae", "network": {"id": "17e72c3b-603f-4305-96ac-89a9b4c9c7e9", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-123501993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7db5465cb5db465b92dfa570572b6e80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c7422d2-df", "ovs_interfaceid": "2c7422d2-dfef-4010-a99a-60c510c54a3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.978583] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.979336] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e/9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 913.979891] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-105b6912-6fa4-4d72-8b61-8c15d9fec0af {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.988487] env[62692]: DEBUG oslo_vmware.api [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Waiting for the task: (returnval){ [ 913.988487] env[62692]: value = "task-1141319" [ 913.988487] env[62692]: _type = "Task" [ 913.988487] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.001231] env[62692]: DEBUG oslo_vmware.api [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141319, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.041612] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 914.041879] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 914.042062] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Deleting the datastore file [datastore2] ae44f074-a5c8-4259-99c6-9ce290b1570b {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 914.042345] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2e38e346-4b9e-4b0f-b95a-dde074fc3856 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.049855] env[62692]: DEBUG oslo_vmware.api [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 914.049855] env[62692]: value = "task-1141320" [ 914.049855] env[62692]: _type = "Task" [ 914.049855] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.058736] env[62692]: DEBUG oslo_vmware.api [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1141320, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.147492] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Acquiring lock "628cc505-3edf-4066-91be-da009ebcf219" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.148043] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Lock "628cc505-3edf-4066-91be-da009ebcf219" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.148366] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Acquiring lock "628cc505-3edf-4066-91be-da009ebcf219-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.148584] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Lock "628cc505-3edf-4066-91be-da009ebcf219-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.148763] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Lock "628cc505-3edf-4066-91be-da009ebcf219-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.156316] env[62692]: INFO nova.compute.manager [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Terminating instance [ 914.156624] env[62692]: DEBUG nova.compute.manager [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 914.156891] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 914.157978] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-033d3888-d52b-474e-9888-2f86d0f0b1e1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.166919] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 914.167246] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-245f597b-001e-4ce2-8243-6d4f2d787f6b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.176450] env[62692]: DEBUG oslo_vmware.api [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Waiting for the task: (returnval){ [ 914.176450] env[62692]: value = "task-1141321" [ 914.176450] env[62692]: _type = "Task" [ 914.176450] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.188818] env[62692]: DEBUG oslo_vmware.api [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1141321, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.237026] env[62692]: DEBUG oslo_vmware.api [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141313, 'name': Rename_Task, 'duration_secs': 0.15404} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.237363] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 914.237624] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-37d837b5-4223-4d7f-8e69-6b136466dbd6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.248525] env[62692]: DEBUG oslo_vmware.api [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 914.248525] env[62692]: value = "task-1141322" [ 914.248525] env[62692]: _type = "Task" [ 914.248525] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.260583] env[62692]: DEBUG oslo_vmware.api [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141322, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.361230] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.621s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.361539] env[62692]: DEBUG nova.compute.manager [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 914.364442] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.434s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.364718] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.367520] env[62692]: DEBUG oslo_concurrency.lockutils [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.812s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.369099] env[62692]: INFO nova.compute.claims [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 914.401161] env[62692]: INFO nova.scheduler.client.report [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Deleted allocations for instance 27583cb4-0478-44d8-8970-478682271513 [ 914.421638] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141316, 'name': CreateVM_Task, 'duration_secs': 0.414754} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.425471] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 914.426034] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141317, 'name': CreateVM_Task, 'duration_secs': 0.428871} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.426465] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.426630] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.427110] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 914.427407] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 914.427746] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d1624ff-5812-4d71-b651-0e59b5fe6699 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.430096] env[62692]: DEBUG oslo_concurrency.lockutils [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/427c2d72-be26-42e4-9f8c-ca628806da41" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.430264] env[62692]: DEBUG oslo_concurrency.lockutils [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquired lock "[datastore1] devstack-image-cache_base/427c2d72-be26-42e4-9f8c-ca628806da41" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.430619] env[62692]: DEBUG oslo_concurrency.lockutils [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/427c2d72-be26-42e4-9f8c-ca628806da41" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 914.431194] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01da8a30-38c0-419a-8c5b-fe2e5745a4a0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.440492] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 914.440492] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52ca5b44-b241-c30b-7eb0-28be9f59ed69" [ 914.440492] env[62692]: _type = "Task" [ 914.440492] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.442172] env[62692]: DEBUG oslo_vmware.api [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 914.442172] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]520b1cc1-de00-8fea-42c1-f4377819c93b" [ 914.442172] env[62692]: _type = "Task" [ 914.442172] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.457213] env[62692]: DEBUG oslo_vmware.api [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]520b1cc1-de00-8fea-42c1-f4377819c93b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.463141] env[62692]: DEBUG oslo_concurrency.lockutils [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Releasing lock "[datastore1] devstack-image-cache_base/427c2d72-be26-42e4-9f8c-ca628806da41" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.463141] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Processing image 427c2d72-be26-42e4-9f8c-ca628806da41 {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 914.463141] env[62692]: DEBUG oslo_concurrency.lockutils [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/427c2d72-be26-42e4-9f8c-ca628806da41/427c2d72-be26-42e4-9f8c-ca628806da41.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.463141] env[62692]: DEBUG oslo_concurrency.lockutils [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquired lock "[datastore1] devstack-image-cache_base/427c2d72-be26-42e4-9f8c-ca628806da41/427c2d72-be26-42e4-9f8c-ca628806da41.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.463414] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 914.465040] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-191e32b2-7159-4d96-9726-9f33169d1be5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.480700] env[62692]: DEBUG oslo_concurrency.lockutils [req-5d2e74de-296d-4f78-b225-ccb79f971051 req-288b3751-914a-4ca4-aa0e-9c78855d2d01 service nova] Releasing lock "refresh_cache-3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.481063] env[62692]: DEBUG nova.compute.manager [req-5d2e74de-296d-4f78-b225-ccb79f971051 req-288b3751-914a-4ca4-aa0e-9c78855d2d01 service nova] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Received event network-vif-plugged-6702eb8d-7d7d-4384-8eb4-d2196df920ef {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 914.481396] env[62692]: DEBUG oslo_concurrency.lockutils [req-5d2e74de-296d-4f78-b225-ccb79f971051 req-288b3751-914a-4ca4-aa0e-9c78855d2d01 service nova] Acquiring lock "44699df6-5262-4700-a096-8c0e450323d2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.481617] env[62692]: DEBUG oslo_concurrency.lockutils [req-5d2e74de-296d-4f78-b225-ccb79f971051 req-288b3751-914a-4ca4-aa0e-9c78855d2d01 service nova] Lock "44699df6-5262-4700-a096-8c0e450323d2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.481825] env[62692]: DEBUG oslo_concurrency.lockutils [req-5d2e74de-296d-4f78-b225-ccb79f971051 req-288b3751-914a-4ca4-aa0e-9c78855d2d01 service nova] Lock "44699df6-5262-4700-a096-8c0e450323d2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.482044] env[62692]: DEBUG nova.compute.manager [req-5d2e74de-296d-4f78-b225-ccb79f971051 req-288b3751-914a-4ca4-aa0e-9c78855d2d01 service nova] [instance: 44699df6-5262-4700-a096-8c0e450323d2] No waiting events found dispatching network-vif-plugged-6702eb8d-7d7d-4384-8eb4-d2196df920ef {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 914.482263] env[62692]: WARNING nova.compute.manager [req-5d2e74de-296d-4f78-b225-ccb79f971051 req-288b3751-914a-4ca4-aa0e-9c78855d2d01 service nova] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Received unexpected event network-vif-plugged-6702eb8d-7d7d-4384-8eb4-d2196df920ef for instance with vm_state building and task_state spawning. [ 914.482470] env[62692]: DEBUG nova.compute.manager [req-5d2e74de-296d-4f78-b225-ccb79f971051 req-288b3751-914a-4ca4-aa0e-9c78855d2d01 service nova] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Received event network-changed-6702eb8d-7d7d-4384-8eb4-d2196df920ef {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 914.482678] env[62692]: DEBUG nova.compute.manager [req-5d2e74de-296d-4f78-b225-ccb79f971051 req-288b3751-914a-4ca4-aa0e-9c78855d2d01 service nova] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Refreshing instance network info cache due to event network-changed-6702eb8d-7d7d-4384-8eb4-d2196df920ef. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 914.482883] env[62692]: DEBUG oslo_concurrency.lockutils [req-5d2e74de-296d-4f78-b225-ccb79f971051 req-288b3751-914a-4ca4-aa0e-9c78855d2d01 service nova] Acquiring lock "refresh_cache-44699df6-5262-4700-a096-8c0e450323d2" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.483077] env[62692]: DEBUG oslo_concurrency.lockutils [req-5d2e74de-296d-4f78-b225-ccb79f971051 req-288b3751-914a-4ca4-aa0e-9c78855d2d01 service nova] Acquired lock "refresh_cache-44699df6-5262-4700-a096-8c0e450323d2" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.483292] env[62692]: DEBUG nova.network.neutron [req-5d2e74de-296d-4f78-b225-ccb79f971051 req-288b3751-914a-4ca4-aa0e-9c78855d2d01 service nova] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Refreshing network info cache for port 6702eb8d-7d7d-4384-8eb4-d2196df920ef {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 914.490187] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 914.490513] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 914.496156] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e91867e9-3f93-4f09-865b-498120af186d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.509613] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 914.509613] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52ccdcd6-f0ca-d778-bf96-7ab93dec206d" [ 914.509613] env[62692]: _type = "Task" [ 914.509613] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.515051] env[62692]: DEBUG oslo_vmware.api [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141319, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.531193] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52ccdcd6-f0ca-d778-bf96-7ab93dec206d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.562015] env[62692]: DEBUG oslo_vmware.api [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1141320, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.443861} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.562475] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 914.562719] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 914.562941] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 914.563171] env[62692]: INFO nova.compute.manager [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Took 1.15 seconds to destroy the instance on the hypervisor. [ 914.563762] env[62692]: DEBUG oslo.service.loopingcall [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.563840] env[62692]: DEBUG nova.compute.manager [-] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 914.563944] env[62692]: DEBUG nova.network.neutron [-] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 914.688051] env[62692]: DEBUG oslo_vmware.api [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1141321, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.759776] env[62692]: DEBUG oslo_vmware.api [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141322, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.874978] env[62692]: DEBUG nova.compute.utils [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 914.878216] env[62692]: DEBUG nova.compute.manager [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Not allocating networking since 'none' was specified. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 914.918764] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d06d1ca1-94c5-4513-9e6c-21ba3545cbf1 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "27583cb4-0478-44d8-8970-478682271513" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.042s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.958394] env[62692]: DEBUG oslo_vmware.api [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]520b1cc1-de00-8fea-42c1-f4377819c93b, 'name': SearchDatastore_Task, 'duration_secs': 0.059344} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.958632] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.958870] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 914.960408] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.960408] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.960408] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 914.960408] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dbdcb90c-66fc-4dbf-9d95-7248ea4e5e18 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.970829] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 914.971080] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 914.971864] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93635f03-0aae-4f78-b8da-4a016dd853b1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.977906] env[62692]: DEBUG oslo_vmware.api [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 914.977906] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52c44851-8d5c-b98f-2b7a-a63356c18800" [ 914.977906] env[62692]: _type = "Task" [ 914.977906] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.988486] env[62692]: DEBUG oslo_vmware.api [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52c44851-8d5c-b98f-2b7a-a63356c18800, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.002262] env[62692]: DEBUG oslo_vmware.api [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141319, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.570145} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.002542] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e/9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 915.002815] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 915.002999] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-117707e6-0a5b-4f11-8212-f20cd8f9a133 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.011189] env[62692]: DEBUG oslo_vmware.api [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Waiting for the task: (returnval){ [ 915.011189] env[62692]: value = "task-1141323" [ 915.011189] env[62692]: _type = "Task" [ 915.011189] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.023895] env[62692]: DEBUG oslo_vmware.api [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141323, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.027635] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Preparing fetch location {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 915.027821] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Fetch image to [datastore1] OSTACK_IMG_e0117b00-4a40-46ab-924b-3823de248860/OSTACK_IMG_e0117b00-4a40-46ab-924b-3823de248860.vmdk {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 915.028014] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Downloading stream optimized image 427c2d72-be26-42e4-9f8c-ca628806da41 to [datastore1] OSTACK_IMG_e0117b00-4a40-46ab-924b-3823de248860/OSTACK_IMG_e0117b00-4a40-46ab-924b-3823de248860.vmdk on the data store datastore1 as vApp {{(pid=62692) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 915.028818] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Downloading image file data 427c2d72-be26-42e4-9f8c-ca628806da41 to the ESX as VM named 'OSTACK_IMG_e0117b00-4a40-46ab-924b-3823de248860' {{(pid=62692) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 915.124445] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c4e580d-f4ee-4b4c-9b15-56a1b2ea13e1 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Volume attach. Driver type: vmdk {{(pid=62692) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 915.124591] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c4e580d-f4ee-4b4c-9b15-56a1b2ea13e1 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-248974', 'volume_id': '57d15e0c-bb59-47d1-abf1-252dc8cb294a', 'name': 'volume-57d15e0c-bb59-47d1-abf1-252dc8cb294a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '754b3c65-1e4b-49d2-8980-095d975edb01', 'attached_at': '', 'detached_at': '', 'volume_id': '57d15e0c-bb59-47d1-abf1-252dc8cb294a', 'serial': '57d15e0c-bb59-47d1-abf1-252dc8cb294a'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 915.125837] env[62692]: DEBUG oslo_vmware.rw_handles [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 915.125837] env[62692]: value = "resgroup-9" [ 915.125837] env[62692]: _type = "ResourcePool" [ 915.125837] env[62692]: }. {{(pid=62692) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 915.126690] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e218e7bb-6163-4e2b-b3e2-5649ccf94051 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.129924] env[62692]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-11e4d2f0-2461-41de-ae8b-276c72fed4d6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.169535] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-657c643b-2e58-4db6-bbd9-6ea715298e83 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.172487] env[62692]: DEBUG oslo_vmware.rw_handles [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lease: (returnval){ [ 915.172487] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]526f1fe4-4641-38d6-d535-06d7a2c74c17" [ 915.172487] env[62692]: _type = "HttpNfcLease" [ 915.172487] env[62692]: } obtained for vApp import into resource pool (val){ [ 915.172487] env[62692]: value = "resgroup-9" [ 915.172487] env[62692]: _type = "ResourcePool" [ 915.172487] env[62692]: }. {{(pid=62692) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 915.172821] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the lease: (returnval){ [ 915.172821] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]526f1fe4-4641-38d6-d535-06d7a2c74c17" [ 915.172821] env[62692]: _type = "HttpNfcLease" [ 915.172821] env[62692]: } to be ready. {{(pid=62692) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 915.203186] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c4e580d-f4ee-4b4c-9b15-56a1b2ea13e1 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] volume-57d15e0c-bb59-47d1-abf1-252dc8cb294a/volume-57d15e0c-bb59-47d1-abf1-252dc8cb294a.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 915.207748] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-87246f42-4e1f-4a90-9a48-075e75825ee6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.222279] env[62692]: DEBUG nova.compute.manager [req-a3e7dd8c-041f-456e-a647-6f8537d7b4e3 req-2bd6134b-68ab-4707-8501-183282ad2fe5 service nova] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Received event network-vif-deleted-f6f97326-a148-4785-a942-d2869ddffca3 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 915.222506] env[62692]: INFO nova.compute.manager [req-a3e7dd8c-041f-456e-a647-6f8537d7b4e3 req-2bd6134b-68ab-4707-8501-183282ad2fe5 service nova] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Neutron deleted interface f6f97326-a148-4785-a942-d2869ddffca3; detaching it from the instance and deleting it from the info cache [ 915.222680] env[62692]: DEBUG nova.network.neutron [req-a3e7dd8c-041f-456e-a647-6f8537d7b4e3 req-2bd6134b-68ab-4707-8501-183282ad2fe5 service nova] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.225768] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 915.225768] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]526f1fe4-4641-38d6-d535-06d7a2c74c17" [ 915.225768] env[62692]: _type = "HttpNfcLease" [ 915.225768] env[62692]: } is initializing. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 915.233390] env[62692]: DEBUG oslo_vmware.api [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1141321, 'name': PowerOffVM_Task, 'duration_secs': 0.52111} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.235227] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 915.235447] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 915.236671] env[62692]: DEBUG oslo_vmware.api [None req-5c4e580d-f4ee-4b4c-9b15-56a1b2ea13e1 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 915.236671] env[62692]: value = "task-1141325" [ 915.236671] env[62692]: _type = "Task" [ 915.236671] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.236671] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e33d1ad6-aec0-4592-8dc5-c0cb62a954f6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.247577] env[62692]: DEBUG oslo_vmware.api [None req-5c4e580d-f4ee-4b4c-9b15-56a1b2ea13e1 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141325, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.262918] env[62692]: DEBUG oslo_vmware.api [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141322, 'name': PowerOnVM_Task, 'duration_secs': 0.644255} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.263355] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 915.263599] env[62692]: INFO nova.compute.manager [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Took 8.13 seconds to spawn the instance on the hypervisor. [ 915.263795] env[62692]: DEBUG nova.compute.manager [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 915.264924] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca89dddc-0724-4c0f-923c-42925119f952 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.340707] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 915.340962] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 915.341173] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Deleting the datastore file [datastore2] 628cc505-3edf-4066-91be-da009ebcf219 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 915.341467] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4a67843e-0650-4916-bd09-305fe7ecc252 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.349930] env[62692]: DEBUG oslo_vmware.api [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Waiting for the task: (returnval){ [ 915.349930] env[62692]: value = "task-1141327" [ 915.349930] env[62692]: _type = "Task" [ 915.349930] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.356208] env[62692]: DEBUG nova.network.neutron [req-5d2e74de-296d-4f78-b225-ccb79f971051 req-288b3751-914a-4ca4-aa0e-9c78855d2d01 service nova] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Updated VIF entry in instance network info cache for port 6702eb8d-7d7d-4384-8eb4-d2196df920ef. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 915.356575] env[62692]: DEBUG nova.network.neutron [req-5d2e74de-296d-4f78-b225-ccb79f971051 req-288b3751-914a-4ca4-aa0e-9c78855d2d01 service nova] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Updating instance_info_cache with network_info: [{"id": "6702eb8d-7d7d-4384-8eb4-d2196df920ef", "address": "fa:16:3e:ba:19:e9", "network": {"id": "2c82845f-101b-48c4-bd0b-4704573a3615", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1972972124-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b0b2e09a06a4fc7a00a03f72989ffed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16f09e8c-5240-4839-80cc-62ec29700bd2", "external-id": "nsx-vlan-transportzone-720", "segmentation_id": 720, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6702eb8d-7d", "ovs_interfaceid": "6702eb8d-7d7d-4384-8eb4-d2196df920ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.365925] env[62692]: DEBUG oslo_vmware.api [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1141327, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.378803] env[62692]: DEBUG nova.compute.manager [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 915.424924] env[62692]: DEBUG nova.network.neutron [-] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.489234] env[62692]: DEBUG oslo_vmware.api [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52c44851-8d5c-b98f-2b7a-a63356c18800, 'name': SearchDatastore_Task, 'duration_secs': 0.012231} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.490119] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06f94a18-6d9e-4160-bf33-c9be5e1de8df {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.500342] env[62692]: DEBUG oslo_vmware.api [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 915.500342] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5248b796-a638-93ae-6d78-b59fc7409502" [ 915.500342] env[62692]: _type = "Task" [ 915.500342] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.513429] env[62692]: DEBUG oslo_vmware.api [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5248b796-a638-93ae-6d78-b59fc7409502, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.523196] env[62692]: DEBUG oslo_vmware.api [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141323, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079408} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.523480] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 915.524093] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15160625-9338-4352-a3a4-75b544380842 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.555364] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e/9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 915.555711] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd483933-3b05-4a5a-986e-aadf7652e366 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.584291] env[62692]: DEBUG oslo_vmware.api [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Waiting for the task: (returnval){ [ 915.584291] env[62692]: value = "task-1141328" [ 915.584291] env[62692]: _type = "Task" [ 915.584291] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.605301] env[62692]: DEBUG oslo_vmware.api [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141328, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.680874] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 915.680874] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]526f1fe4-4641-38d6-d535-06d7a2c74c17" [ 915.680874] env[62692]: _type = "HttpNfcLease" [ 915.680874] env[62692]: } is initializing. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 915.727188] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d58de596-0c9f-4ab2-8ef9-9e7477f192bc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.744433] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55143272-9149-47a2-a226-e2089820cec3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.764139] env[62692]: DEBUG oslo_vmware.api [None req-5c4e580d-f4ee-4b4c-9b15-56a1b2ea13e1 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141325, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.794606] env[62692]: DEBUG nova.compute.manager [req-a3e7dd8c-041f-456e-a647-6f8537d7b4e3 req-2bd6134b-68ab-4707-8501-183282ad2fe5 service nova] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Detach interface failed, port_id=f6f97326-a148-4785-a942-d2869ddffca3, reason: Instance ae44f074-a5c8-4259-99c6-9ce290b1570b could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 915.800712] env[62692]: INFO nova.compute.manager [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Took 44.67 seconds to build instance. [ 915.860256] env[62692]: DEBUG oslo_concurrency.lockutils [req-5d2e74de-296d-4f78-b225-ccb79f971051 req-288b3751-914a-4ca4-aa0e-9c78855d2d01 service nova] Releasing lock "refresh_cache-44699df6-5262-4700-a096-8c0e450323d2" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.864829] env[62692]: DEBUG oslo_vmware.api [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Task: {'id': task-1141327, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.475567} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.865476] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 915.865476] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 915.865595] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 915.865727] env[62692]: INFO nova.compute.manager [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Took 1.71 seconds to destroy the instance on the hypervisor. [ 915.865979] env[62692]: DEBUG oslo.service.loopingcall [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.866209] env[62692]: DEBUG nova.compute.manager [-] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 915.866287] env[62692]: DEBUG nova.network.neutron [-] [instance: 628cc505-3edf-4066-91be-da009ebcf219] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 915.929598] env[62692]: INFO nova.compute.manager [-] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Took 1.37 seconds to deallocate network for instance. [ 915.936444] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9649e368-ce91-4e98-8cda-321234650546 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.945381] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6357e76-9ac6-4105-a749-4e9b625764a0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.982484] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cca295c-66cb-4b6f-bc78-a885591513e0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.991871] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb0c7d31-c6dc-4d4f-8780-e40fb5a55fc7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.008602] env[62692]: DEBUG nova.compute.provider_tree [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.019826] env[62692]: DEBUG oslo_vmware.api [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5248b796-a638-93ae-6d78-b59fc7409502, 'name': SearchDatastore_Task, 'duration_secs': 0.031682} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.020765] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.021071] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957/fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 916.021360] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-937bf08d-54e8-4ecd-8246-91c64ef0a2fb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.030511] env[62692]: DEBUG oslo_vmware.api [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 916.030511] env[62692]: value = "task-1141329" [ 916.030511] env[62692]: _type = "Task" [ 916.030511] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.040826] env[62692]: DEBUG oslo_vmware.api [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141329, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.097814] env[62692]: DEBUG oslo_vmware.api [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141328, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.181506] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 916.181506] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]526f1fe4-4641-38d6-d535-06d7a2c74c17" [ 916.181506] env[62692]: _type = "HttpNfcLease" [ 916.181506] env[62692]: } is initializing. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 916.252158] env[62692]: DEBUG oslo_vmware.api [None req-5c4e580d-f4ee-4b4c-9b15-56a1b2ea13e1 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141325, 'name': ReconfigVM_Task, 'duration_secs': 0.91095} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.252474] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c4e580d-f4ee-4b4c-9b15-56a1b2ea13e1 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Reconfigured VM instance instance-0000003b to attach disk [datastore1] volume-57d15e0c-bb59-47d1-abf1-252dc8cb294a/volume-57d15e0c-bb59-47d1-abf1-252dc8cb294a.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 916.257669] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a85c981c-ff53-434d-8c31-29b890538f0f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.280878] env[62692]: DEBUG oslo_vmware.api [None req-5c4e580d-f4ee-4b4c-9b15-56a1b2ea13e1 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 916.280878] env[62692]: value = "task-1141330" [ 916.280878] env[62692]: _type = "Task" [ 916.280878] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.292205] env[62692]: DEBUG oslo_vmware.api [None req-5c4e580d-f4ee-4b4c-9b15-56a1b2ea13e1 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141330, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.304150] env[62692]: DEBUG oslo_concurrency.lockutils [None req-06d631cd-a440-468e-b98c-691cf355a965 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "3fd2959b-80cb-470c-8c2a-40b7630458ed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.182s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.391609] env[62692]: DEBUG nova.compute.manager [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 916.425685] env[62692]: DEBUG nova.virt.hardware [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 916.426055] env[62692]: DEBUG nova.virt.hardware [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 916.426259] env[62692]: DEBUG nova.virt.hardware [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 916.426492] env[62692]: DEBUG nova.virt.hardware [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 916.426597] env[62692]: DEBUG nova.virt.hardware [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 916.426824] env[62692]: DEBUG nova.virt.hardware [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 916.427125] env[62692]: DEBUG nova.virt.hardware [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 916.427307] env[62692]: DEBUG nova.virt.hardware [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 916.427484] env[62692]: DEBUG nova.virt.hardware [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 916.427658] env[62692]: DEBUG nova.virt.hardware [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 916.427871] env[62692]: DEBUG nova.virt.hardware [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 916.428936] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acfcc3e1-0bba-453b-ad08-9fd47d4043f2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.442624] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.444499] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eb547fe-593d-44d3-838a-8ebf27b9d1c4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.465789] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Instance VIF info [] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 916.473996] env[62692]: DEBUG oslo.service.loopingcall [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 916.474546] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fac426b6-747a-455e-85a1-44e71b432180] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 916.474881] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bb9b7007-1422-4f67-90a2-c70c8f314dae {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.502821] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 916.502821] env[62692]: value = "task-1141331" [ 916.502821] env[62692]: _type = "Task" [ 916.502821] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.515459] env[62692]: DEBUG nova.scheduler.client.report [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 916.527968] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141331, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.547789] env[62692]: DEBUG oslo_vmware.api [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141329, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.596857] env[62692]: DEBUG oslo_vmware.api [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141328, 'name': ReconfigVM_Task, 'duration_secs': 0.588992} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.597208] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e/9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 916.598041] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f7a7c9ba-6022-4193-b862-0a32d629dce0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.608848] env[62692]: DEBUG oslo_vmware.api [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Waiting for the task: (returnval){ [ 916.608848] env[62692]: value = "task-1141332" [ 916.608848] env[62692]: _type = "Task" [ 916.608848] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.620034] env[62692]: DEBUG oslo_vmware.api [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141332, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.683330] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 916.683330] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]526f1fe4-4641-38d6-d535-06d7a2c74c17" [ 916.683330] env[62692]: _type = "HttpNfcLease" [ 916.683330] env[62692]: } is ready. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 916.683694] env[62692]: DEBUG oslo_vmware.rw_handles [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 916.683694] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]526f1fe4-4641-38d6-d535-06d7a2c74c17" [ 916.683694] env[62692]: _type = "HttpNfcLease" [ 916.683694] env[62692]: }. {{(pid=62692) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 916.685173] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fe145e7-ae54-45de-94dd-02280f3dc071 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.694303] env[62692]: DEBUG oslo_vmware.rw_handles [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52295e37-1086-2d31-7d78-dbc0a80a4403/disk-0.vmdk from lease info. {{(pid=62692) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 916.694576] env[62692]: DEBUG oslo_vmware.rw_handles [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52295e37-1086-2d31-7d78-dbc0a80a4403/disk-0.vmdk. {{(pid=62692) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 916.766817] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-5404a689-9481-473c-b799-b2cb478eedc7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.806875] env[62692]: DEBUG oslo_vmware.api [None req-5c4e580d-f4ee-4b4c-9b15-56a1b2ea13e1 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141330, 'name': ReconfigVM_Task, 'duration_secs': 0.173959} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.807249] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c4e580d-f4ee-4b4c-9b15-56a1b2ea13e1 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-248974', 'volume_id': '57d15e0c-bb59-47d1-abf1-252dc8cb294a', 'name': 'volume-57d15e0c-bb59-47d1-abf1-252dc8cb294a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '754b3c65-1e4b-49d2-8980-095d975edb01', 'attached_at': '', 'detached_at': '', 'volume_id': '57d15e0c-bb59-47d1-abf1-252dc8cb294a', 'serial': '57d15e0c-bb59-47d1-abf1-252dc8cb294a'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 916.808845] env[62692]: DEBUG nova.compute.manager [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 917.020241] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141331, 'name': CreateVM_Task, 'duration_secs': 0.427298} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.022868] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fac426b6-747a-455e-85a1-44e71b432180] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 917.023497] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.023665] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.024119] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 917.024506] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a4e01c6-e5a5-48dd-b421-f96c224e0f53 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.030125] env[62692]: DEBUG oslo_concurrency.lockutils [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.663s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.030836] env[62692]: DEBUG nova.compute.manager [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 917.037491] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.591s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.040404] env[62692]: INFO nova.compute.claims [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 917.042366] env[62692]: DEBUG oslo_vmware.api [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 917.042366] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52378949-ae91-ac35-2179-a4dfa71ea0c0" [ 917.042366] env[62692]: _type = "Task" [ 917.042366] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.061624] env[62692]: DEBUG oslo_vmware.api [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52378949-ae91-ac35-2179-a4dfa71ea0c0, 'name': SearchDatastore_Task, 'duration_secs': 0.014356} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.066590] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.066906] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 917.067167] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.067323] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.067509] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 917.067888] env[62692]: DEBUG oslo_vmware.api [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141329, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.577537} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.069643] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-50eead22-fa13-40bc-9f3e-3aa331d613fd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.071885] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957/fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 917.072127] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 917.072449] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e221bcd6-a88b-4333-9915-2f63146bc435 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.082615] env[62692]: DEBUG oslo_vmware.api [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 917.082615] env[62692]: value = "task-1141333" [ 917.082615] env[62692]: _type = "Task" [ 917.082615] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.083043] env[62692]: DEBUG nova.network.neutron [-] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.090196] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 917.090477] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 917.092962] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ee8df30-0e49-4ab4-8e6c-795c2122139b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.104356] env[62692]: DEBUG oslo_vmware.api [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141333, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.107632] env[62692]: DEBUG oslo_vmware.api [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 917.107632] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]522fa3d4-24ab-ac7d-a248-dc75d8e3b57f" [ 917.107632] env[62692]: _type = "Task" [ 917.107632] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.124373] env[62692]: DEBUG oslo_vmware.api [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]522fa3d4-24ab-ac7d-a248-dc75d8e3b57f, 'name': SearchDatastore_Task, 'duration_secs': 0.013233} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.130803] env[62692]: DEBUG oslo_vmware.api [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141332, 'name': Rename_Task, 'duration_secs': 0.176653} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.131122] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1846b3de-2ded-48d2-8bd8-ae138da62463 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.133858] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 917.135457] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f5737898-9e24-4a65-b5fb-d163ea8d040c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.142661] env[62692]: DEBUG oslo_vmware.api [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 917.142661] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]527e852e-f7d6-f218-8513-a60f46b63eef" [ 917.142661] env[62692]: _type = "Task" [ 917.142661] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.148830] env[62692]: DEBUG oslo_vmware.api [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Waiting for the task: (returnval){ [ 917.148830] env[62692]: value = "task-1141334" [ 917.148830] env[62692]: _type = "Task" [ 917.148830] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.157432] env[62692]: DEBUG oslo_vmware.api [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]527e852e-f7d6-f218-8513-a60f46b63eef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.164904] env[62692]: DEBUG oslo_vmware.api [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141334, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.222734] env[62692]: DEBUG nova.compute.manager [req-6516d2df-6ede-4ce8-841c-a68fd112000a req-533158a0-9734-4317-a183-7636929395f6 service nova] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Received event network-vif-deleted-30e027a7-86b8-4669-91fe-1dbd40ae54a5 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 917.332971] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.535422] env[62692]: DEBUG oslo_vmware.rw_handles [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Completed reading data from the image iterator. {{(pid=62692) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 917.535717] env[62692]: DEBUG oslo_vmware.rw_handles [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52295e37-1086-2d31-7d78-dbc0a80a4403/disk-0.vmdk. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 917.536670] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d99069cf-d489-47a0-a205-24bf1f7569f6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.547984] env[62692]: DEBUG nova.compute.utils [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 917.549277] env[62692]: DEBUG oslo_vmware.rw_handles [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52295e37-1086-2d31-7d78-dbc0a80a4403/disk-0.vmdk is in state: ready. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 917.549496] env[62692]: DEBUG oslo_vmware.rw_handles [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52295e37-1086-2d31-7d78-dbc0a80a4403/disk-0.vmdk. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 917.550089] env[62692]: DEBUG nova.compute.manager [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 917.550464] env[62692]: DEBUG nova.network.neutron [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 917.552711] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-3f2e1216-0545-4747-b64d-cc3e3dc23190 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.599018] env[62692]: INFO nova.compute.manager [-] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Took 1.73 seconds to deallocate network for instance. [ 917.599018] env[62692]: DEBUG oslo_vmware.api [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141333, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.263159} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.600643] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 917.601739] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a17b563-c234-4d5e-a8d4-ee1044871d91 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.632856] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957/fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 917.635128] env[62692]: DEBUG nova.policy [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9e9f256c1ebf4b1ba966edb7b1df5f4f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '266c39fe585d41079d0278ecd1d1c6a4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 917.637996] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f61a8948-f1d2-40bc-b495-c79b48962872 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.675374] env[62692]: DEBUG oslo_vmware.api [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]527e852e-f7d6-f218-8513-a60f46b63eef, 'name': SearchDatastore_Task, 'duration_secs': 0.013825} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.680770] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.681176] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] fac426b6-747a-455e-85a1-44e71b432180/fac426b6-747a-455e-85a1-44e71b432180.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 917.681645] env[62692]: DEBUG oslo_vmware.api [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 917.681645] env[62692]: value = "task-1141335" [ 917.681645] env[62692]: _type = "Task" [ 917.681645] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.682327] env[62692]: DEBUG oslo_vmware.api [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141334, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.682769] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c433d6a6-d806-49e0-bc80-467deb7016fb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.698287] env[62692]: DEBUG oslo_vmware.api [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141335, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.698548] env[62692]: DEBUG oslo_vmware.api [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 917.698548] env[62692]: value = "task-1141336" [ 917.698548] env[62692]: _type = "Task" [ 917.698548] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.712666] env[62692]: DEBUG oslo_vmware.api [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141336, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.755515] env[62692]: DEBUG oslo_concurrency.lockutils [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "668fb0ea-7eed-4198-943c-8f916ec11368" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.755918] env[62692]: DEBUG oslo_concurrency.lockutils [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "668fb0ea-7eed-4198-943c-8f916ec11368" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.830350] env[62692]: DEBUG oslo_vmware.rw_handles [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52295e37-1086-2d31-7d78-dbc0a80a4403/disk-0.vmdk. {{(pid=62692) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 917.830822] env[62692]: INFO nova.virt.vmwareapi.images [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Downloaded image file data 427c2d72-be26-42e4-9f8c-ca628806da41 [ 917.831672] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d40f4c8-d05c-4aff-b20b-42f3819421ca {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.850469] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2edb4870-0de4-4b21-a544-c19cd29dc9a9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.858891] env[62692]: DEBUG nova.objects.instance [None req-5c4e580d-f4ee-4b4c-9b15-56a1b2ea13e1 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lazy-loading 'flavor' on Instance uuid 754b3c65-1e4b-49d2-8980-095d975edb01 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 917.872428] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ea85a1e4-edbb-4f66-9d9a-1923c1786ac2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "3fd2959b-80cb-470c-8c2a-40b7630458ed" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.872823] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ea85a1e4-edbb-4f66-9d9a-1923c1786ac2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "3fd2959b-80cb-470c-8c2a-40b7630458ed" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.887384] env[62692]: INFO nova.virt.vmwareapi.images [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] The imported VM was unregistered [ 917.890346] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Caching image {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 917.890694] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Creating directory with path [datastore1] devstack-image-cache_base/427c2d72-be26-42e4-9f8c-ca628806da41 {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 917.891316] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-29bdd05d-7fc6-42b1-92d3-a6b24c311a25 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.909164] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Created directory with path [datastore1] devstack-image-cache_base/427c2d72-be26-42e4-9f8c-ca628806da41 {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 917.909473] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_e0117b00-4a40-46ab-924b-3823de248860/OSTACK_IMG_e0117b00-4a40-46ab-924b-3823de248860.vmdk to [datastore1] devstack-image-cache_base/427c2d72-be26-42e4-9f8c-ca628806da41/427c2d72-be26-42e4-9f8c-ca628806da41.vmdk. {{(pid=62692) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 917.909796] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-7b566dbd-fe73-42f6-8f0b-f7485eadd30e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.918352] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 917.918352] env[62692]: value = "task-1141338" [ 917.918352] env[62692]: _type = "Task" [ 917.918352] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.927854] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141338, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.057984] env[62692]: DEBUG nova.compute.manager [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 918.101824] env[62692]: DEBUG nova.network.neutron [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Successfully created port: 0047b080-5217-4718-a134-a87f4556102b {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 918.112186] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.169017] env[62692]: DEBUG oslo_vmware.api [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141334, 'name': PowerOnVM_Task, 'duration_secs': 0.668104} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.169470] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 918.169757] env[62692]: INFO nova.compute.manager [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Took 13.65 seconds to spawn the instance on the hypervisor. [ 918.169888] env[62692]: DEBUG nova.compute.manager [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 918.170848] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a96abbe-7960-4dc2-9945-414265334f0d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.201247] env[62692]: DEBUG oslo_vmware.api [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141335, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.214598] env[62692]: DEBUG oslo_vmware.api [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141336, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.365015] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5c4e580d-f4ee-4b4c-9b15-56a1b2ea13e1 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "754b3c65-1e4b-49d2-8980-095d975edb01" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.878s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.377252] env[62692]: DEBUG nova.compute.utils [None req-ea85a1e4-edbb-4f66-9d9a-1923c1786ac2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 918.431046] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141338, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.569028] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53e4005d-3a3d-4f2f-a192-23582c2e19df {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.576983] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f715c0c-6c6c-4ef6-a8e0-74bdafebf6a3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.611989] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4abf666-e87a-4305-8903-7be0aa58cb49 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.626065] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab021964-c6a7-4b4a-b796-a9635ddb6ab2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.645351] env[62692]: DEBUG nova.compute.provider_tree [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 918.695392] env[62692]: INFO nova.compute.manager [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Took 59.43 seconds to build instance. [ 918.700501] env[62692]: DEBUG oslo_vmware.api [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141335, 'name': ReconfigVM_Task, 'duration_secs': 0.596731} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.700831] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Reconfigured VM instance instance-00000045 to attach disk [datastore2] fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957/fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 918.701609] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e549ce3e-b51e-4a4b-88a2-63a70c33f385 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.716823] env[62692]: DEBUG oslo_vmware.api [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 918.716823] env[62692]: value = "task-1141339" [ 918.716823] env[62692]: _type = "Task" [ 918.716823] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.721913] env[62692]: DEBUG oslo_vmware.api [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141336, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.754651} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.725611] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] fac426b6-747a-455e-85a1-44e71b432180/fac426b6-747a-455e-85a1-44e71b432180.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 918.725910] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 918.726533] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-10ddd387-3575-4aa9-9237-3f79f1aa8628 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.745160] env[62692]: DEBUG oslo_vmware.api [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141339, 'name': Rename_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.747220] env[62692]: DEBUG oslo_vmware.api [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 918.747220] env[62692]: value = "task-1141340" [ 918.747220] env[62692]: _type = "Task" [ 918.747220] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.760310] env[62692]: DEBUG oslo_vmware.api [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141340, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.881000] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ea85a1e4-edbb-4f66-9d9a-1923c1786ac2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "3fd2959b-80cb-470c-8c2a-40b7630458ed" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.929558] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141338, 'name': MoveVirtualDisk_Task} progress is 18%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.075794] env[62692]: DEBUG nova.compute.manager [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 919.107022] env[62692]: DEBUG nova.virt.hardware [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 919.107022] env[62692]: DEBUG nova.virt.hardware [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 919.107022] env[62692]: DEBUG nova.virt.hardware [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 919.107022] env[62692]: DEBUG nova.virt.hardware [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 919.107022] env[62692]: DEBUG nova.virt.hardware [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 919.107022] env[62692]: DEBUG nova.virt.hardware [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 919.107618] env[62692]: DEBUG nova.virt.hardware [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 919.108192] env[62692]: DEBUG nova.virt.hardware [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 919.109043] env[62692]: DEBUG nova.virt.hardware [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 919.109043] env[62692]: DEBUG nova.virt.hardware [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 919.109580] env[62692]: DEBUG nova.virt.hardware [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 919.110888] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88fed2cb-7d43-4560-93ce-ac983273fe69 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.123191] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f27409c0-4599-4fca-86f5-e2e352dd38b1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.149237] env[62692]: DEBUG nova.scheduler.client.report [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 919.197762] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ae3da79-c59e-4471-9ceb-4d74d7a720ec tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Lock "9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.946s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.236135] env[62692]: DEBUG oslo_vmware.api [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141339, 'name': Rename_Task, 'duration_secs': 0.300675} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.236135] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 919.237615] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7f8904ec-4269-4a95-bad2-5573dae56e76 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.247945] env[62692]: DEBUG oslo_vmware.api [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 919.247945] env[62692]: value = "task-1141341" [ 919.247945] env[62692]: _type = "Task" [ 919.247945] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.262958] env[62692]: DEBUG oslo_vmware.api [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141341, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.266657] env[62692]: DEBUG oslo_vmware.api [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141340, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.099985} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.267030] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 919.267915] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60a004cb-b398-46d9-9c0a-e85239ccb84f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.297233] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] fac426b6-747a-455e-85a1-44e71b432180/fac426b6-747a-455e-85a1-44e71b432180.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 919.297233] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-202af177-925e-4ef9-b4ea-4e43473f365e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.322324] env[62692]: DEBUG oslo_vmware.api [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 919.322324] env[62692]: value = "task-1141342" [ 919.322324] env[62692]: _type = "Task" [ 919.322324] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.336049] env[62692]: DEBUG oslo_vmware.api [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141342, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.435950] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141338, 'name': MoveVirtualDisk_Task} progress is 38%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.657271] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.620s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.657922] env[62692]: DEBUG nova.compute.manager [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 919.664009] env[62692]: DEBUG oslo_concurrency.lockutils [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.102s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.669038] env[62692]: INFO nova.compute.claims [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 919.701020] env[62692]: DEBUG nova.compute.manager [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 919.766233] env[62692]: DEBUG oslo_vmware.api [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141341, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.840704] env[62692]: DEBUG oslo_vmware.api [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141342, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.939507] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141338, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.940692] env[62692]: DEBUG nova.network.neutron [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Successfully updated port: 0047b080-5217-4718-a134-a87f4556102b {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 919.960585] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ea85a1e4-edbb-4f66-9d9a-1923c1786ac2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "3fd2959b-80cb-470c-8c2a-40b7630458ed" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.960960] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ea85a1e4-edbb-4f66-9d9a-1923c1786ac2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "3fd2959b-80cb-470c-8c2a-40b7630458ed" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.961229] env[62692]: INFO nova.compute.manager [None req-ea85a1e4-edbb-4f66-9d9a-1923c1786ac2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Attaching volume 14bc011a-54f2-4452-95c6-b9a4f2020742 to /dev/sdb [ 919.965927] env[62692]: DEBUG nova.compute.manager [req-075d4624-ad9c-4d28-a2d8-3bc9874e0bf8 req-c03919b6-69cc-4fe3-b9ca-9b6b498f367b service nova] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Received event network-vif-plugged-0047b080-5217-4718-a134-a87f4556102b {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 919.966171] env[62692]: DEBUG oslo_concurrency.lockutils [req-075d4624-ad9c-4d28-a2d8-3bc9874e0bf8 req-c03919b6-69cc-4fe3-b9ca-9b6b498f367b service nova] Acquiring lock "0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.966387] env[62692]: DEBUG oslo_concurrency.lockutils [req-075d4624-ad9c-4d28-a2d8-3bc9874e0bf8 req-c03919b6-69cc-4fe3-b9ca-9b6b498f367b service nova] Lock "0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.966633] env[62692]: DEBUG oslo_concurrency.lockutils [req-075d4624-ad9c-4d28-a2d8-3bc9874e0bf8 req-c03919b6-69cc-4fe3-b9ca-9b6b498f367b service nova] Lock "0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.966854] env[62692]: DEBUG nova.compute.manager [req-075d4624-ad9c-4d28-a2d8-3bc9874e0bf8 req-c03919b6-69cc-4fe3-b9ca-9b6b498f367b service nova] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] No waiting events found dispatching network-vif-plugged-0047b080-5217-4718-a134-a87f4556102b {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 919.967046] env[62692]: WARNING nova.compute.manager [req-075d4624-ad9c-4d28-a2d8-3bc9874e0bf8 req-c03919b6-69cc-4fe3-b9ca-9b6b498f367b service nova] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Received unexpected event network-vif-plugged-0047b080-5217-4718-a134-a87f4556102b for instance with vm_state building and task_state spawning. [ 920.008034] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6236063-e517-4c9d-83d4-ab9ee1184a69 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.021026] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84895f5d-2959-46df-a3a8-74d9dc7e1077 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.038686] env[62692]: DEBUG nova.virt.block_device [None req-ea85a1e4-edbb-4f66-9d9a-1923c1786ac2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Updating existing volume attachment record: f47bbf3d-f5ca-4c7a-9231-745859e5a3e1 {{(pid=62692) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 920.064822] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Acquiring lock "9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.065158] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Lock "9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.065385] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Acquiring lock "9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.065574] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Lock "9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.065746] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Lock "9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.068153] env[62692]: INFO nova.compute.manager [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Terminating instance [ 920.070320] env[62692]: DEBUG nova.compute.manager [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 920.070526] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 920.071419] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abdb440b-a98c-4903-ba90-5e95379cd635 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.085023] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 920.085359] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-41e819e9-010f-4df9-ac8c-b7fbe66ab085 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.094370] env[62692]: DEBUG oslo_vmware.api [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Waiting for the task: (returnval){ [ 920.094370] env[62692]: value = "task-1141343" [ 920.094370] env[62692]: _type = "Task" [ 920.094370] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.106980] env[62692]: DEBUG oslo_vmware.api [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141343, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.165694] env[62692]: DEBUG nova.compute.utils [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 920.168636] env[62692]: DEBUG nova.compute.manager [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 920.168636] env[62692]: DEBUG nova.network.neutron [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 920.229295] env[62692]: DEBUG nova.policy [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f2e2ee26163841ddb15bb67520e98394', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e13f8888d5e84eed8247101558e29a56', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 920.262743] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.268962] env[62692]: DEBUG oslo_vmware.api [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141341, 'name': PowerOnVM_Task, 'duration_secs': 0.729791} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.270239] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 920.270239] env[62692]: INFO nova.compute.manager [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Took 6.50 seconds to spawn the instance on the hypervisor. [ 920.270239] env[62692]: DEBUG nova.compute.manager [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 920.271041] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-860dde31-95ff-457e-b68b-5b7e88fa81e1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.337350] env[62692]: DEBUG oslo_vmware.api [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141342, 'name': ReconfigVM_Task, 'duration_secs': 0.674677} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.337615] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Reconfigured VM instance instance-00000046 to attach disk [datastore1] fac426b6-747a-455e-85a1-44e71b432180/fac426b6-747a-455e-85a1-44e71b432180.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 920.338179] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-be5661c6-b9ef-4e35-80e2-e4c9e5aa5f1e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.351505] env[62692]: DEBUG oslo_vmware.api [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 920.351505] env[62692]: value = "task-1141345" [ 920.351505] env[62692]: _type = "Task" [ 920.351505] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.364356] env[62692]: DEBUG oslo_vmware.api [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141345, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.434923] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141338, 'name': MoveVirtualDisk_Task} progress is 80%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.446318] env[62692]: DEBUG oslo_concurrency.lockutils [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "refresh_cache-0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.446318] env[62692]: DEBUG oslo_concurrency.lockutils [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquired lock "refresh_cache-0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.446318] env[62692]: DEBUG nova.network.neutron [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 920.612767] env[62692]: DEBUG oslo_vmware.api [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141343, 'name': PowerOffVM_Task, 'duration_secs': 0.379874} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.612767] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 920.612767] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 920.612767] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7d11775c-5a98-4cf2-a74f-1cc422dab145 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.646831] env[62692]: DEBUG nova.network.neutron [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Successfully created port: 7cf21363-a84c-4404-b293-ccee3ea80385 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 920.671609] env[62692]: DEBUG nova.compute.manager [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 920.790306] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 920.790652] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 920.790866] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Deleting the datastore file [datastore2] 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 920.795245] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-263d8775-04e3-4737-b3da-6552f8d15fef {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.801830] env[62692]: INFO nova.compute.manager [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Took 44.58 seconds to build instance. [ 920.814779] env[62692]: DEBUG oslo_vmware.api [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Waiting for the task: (returnval){ [ 920.814779] env[62692]: value = "task-1141349" [ 920.814779] env[62692]: _type = "Task" [ 920.814779] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.831026] env[62692]: DEBUG oslo_vmware.api [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141349, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.867564] env[62692]: DEBUG oslo_vmware.api [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141345, 'name': Rename_Task, 'duration_secs': 0.349134} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.870782] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 920.872148] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a3909ebb-f61f-452f-a954-03df4e9f196b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.884194] env[62692]: DEBUG oslo_vmware.api [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 920.884194] env[62692]: value = "task-1141350" [ 920.884194] env[62692]: _type = "Task" [ 920.884194] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.898735] env[62692]: DEBUG oslo_vmware.api [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141350, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.911774] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "f3e8be1c-d952-49b7-86a8-d0a6c56ee096" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.912322] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "f3e8be1c-d952-49b7-86a8-d0a6c56ee096" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.936487] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141338, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.007037] env[62692]: DEBUG nova.network.neutron [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 921.185224] env[62692]: DEBUG nova.network.neutron [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Updating instance_info_cache with network_info: [{"id": "0047b080-5217-4718-a134-a87f4556102b", "address": "fa:16:3e:0d:46:60", "network": {"id": "0e06e04a-5213-4d06-b3f1-b2b8f437e6d0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1533469056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "266c39fe585d41079d0278ecd1d1c6a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "678ebbe4-4c53-4eaf-a689-93981310f37d", "external-id": "nsx-vlan-transportzone-443", "segmentation_id": 443, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0047b080-52", "ovs_interfaceid": "0047b080-5217-4718-a134-a87f4556102b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.238357] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86e2f9a4-f846-4d65-8264-c73ede8b6450 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.247722] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3bfe4cc-73ce-42af-8a3c-338fab527dc9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.280997] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bb51401-be83-406a-8ac5-f8b71bb9e254 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.289167] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3469a51c-7d3e-479e-81b5-99782ca033d8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.303590] env[62692]: DEBUG nova.compute.provider_tree [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 921.304974] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f2a0a9da-9e14-46c7-a8c0-b0ae591e1017 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Lock "fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.098s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.325937] env[62692]: DEBUG oslo_vmware.api [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141349, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.399241] env[62692]: DEBUG oslo_vmware.api [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141350, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.433309] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141338, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.358871} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.433646] env[62692]: INFO nova.virt.vmwareapi.ds_util [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_e0117b00-4a40-46ab-924b-3823de248860/OSTACK_IMG_e0117b00-4a40-46ab-924b-3823de248860.vmdk to [datastore1] devstack-image-cache_base/427c2d72-be26-42e4-9f8c-ca628806da41/427c2d72-be26-42e4-9f8c-ca628806da41.vmdk. [ 921.433853] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Cleaning up location [datastore1] OSTACK_IMG_e0117b00-4a40-46ab-924b-3823de248860 {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 921.434030] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_e0117b00-4a40-46ab-924b-3823de248860 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 921.434332] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5216de5e-6b72-4af1-ade4-9422fa1ca811 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.442909] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 921.442909] env[62692]: value = "task-1141351" [ 921.442909] env[62692]: _type = "Task" [ 921.442909] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.201806] env[62692]: DEBUG nova.compute.manager [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 922.204051] env[62692]: DEBUG oslo_concurrency.lockutils [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Releasing lock "refresh_cache-0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.204343] env[62692]: DEBUG nova.compute.manager [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Instance network_info: |[{"id": "0047b080-5217-4718-a134-a87f4556102b", "address": "fa:16:3e:0d:46:60", "network": {"id": "0e06e04a-5213-4d06-b3f1-b2b8f437e6d0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1533469056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "266c39fe585d41079d0278ecd1d1c6a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "678ebbe4-4c53-4eaf-a689-93981310f37d", "external-id": "nsx-vlan-transportzone-443", "segmentation_id": 443, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0047b080-52", "ovs_interfaceid": "0047b080-5217-4718-a134-a87f4556102b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 922.205281] env[62692]: DEBUG nova.scheduler.client.report [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 922.212020] env[62692]: DEBUG nova.compute.manager [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 922.214456] env[62692]: DEBUG nova.compute.manager [req-e2caf092-e2e8-458e-9dcc-6ee4555adc5d req-90df103d-92e2-46b5-93d0-7605fb0be0b6 service nova] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Received event network-changed-0047b080-5217-4718-a134-a87f4556102b {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 922.214634] env[62692]: DEBUG nova.compute.manager [req-e2caf092-e2e8-458e-9dcc-6ee4555adc5d req-90df103d-92e2-46b5-93d0-7605fb0be0b6 service nova] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Refreshing instance network info cache due to event network-changed-0047b080-5217-4718-a134-a87f4556102b. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 922.214847] env[62692]: DEBUG oslo_concurrency.lockutils [req-e2caf092-e2e8-458e-9dcc-6ee4555adc5d req-90df103d-92e2-46b5-93d0-7605fb0be0b6 service nova] Acquiring lock "refresh_cache-0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.215022] env[62692]: DEBUG oslo_concurrency.lockutils [req-e2caf092-e2e8-458e-9dcc-6ee4555adc5d req-90df103d-92e2-46b5-93d0-7605fb0be0b6 service nova] Acquired lock "refresh_cache-0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.215495] env[62692]: DEBUG nova.network.neutron [req-e2caf092-e2e8-458e-9dcc-6ee4555adc5d req-90df103d-92e2-46b5-93d0-7605fb0be0b6 service nova] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Refreshing network info cache for port 0047b080-5217-4718-a134-a87f4556102b {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 922.217307] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0d:46:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '678ebbe4-4c53-4eaf-a689-93981310f37d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0047b080-5217-4718-a134-a87f4556102b', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 922.226388] env[62692]: DEBUG oslo.service.loopingcall [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.238108] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 922.242481] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0bf68185-c092-453c-9a83-fc4171d6b9cc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.272055] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141351, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187364} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.278146] env[62692]: DEBUG nova.virt.hardware [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 922.278532] env[62692]: DEBUG nova.virt.hardware [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 922.279338] env[62692]: DEBUG nova.virt.hardware [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 922.280399] env[62692]: DEBUG nova.virt.hardware [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 922.280635] env[62692]: DEBUG nova.virt.hardware [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 922.280941] env[62692]: DEBUG nova.virt.hardware [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 922.281350] env[62692]: DEBUG nova.virt.hardware [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 922.281567] env[62692]: DEBUG nova.virt.hardware [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 922.281899] env[62692]: DEBUG nova.virt.hardware [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 922.282860] env[62692]: DEBUG nova.virt.hardware [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 922.283095] env[62692]: DEBUG nova.virt.hardware [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 922.296483] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 922.296861] env[62692]: DEBUG oslo_concurrency.lockutils [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Releasing lock "[datastore1] devstack-image-cache_base/427c2d72-be26-42e4-9f8c-ca628806da41/427c2d72-be26-42e4-9f8c-ca628806da41.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.296861] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/427c2d72-be26-42e4-9f8c-ca628806da41/427c2d72-be26-42e4-9f8c-ca628806da41.vmdk to [datastore1] 44699df6-5262-4700-a096-8c0e450323d2/44699df6-5262-4700-a096-8c0e450323d2.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 922.301081] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc738325-f72c-40eb-bbfa-2433a351a9ad {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.305429] env[62692]: DEBUG oslo_vmware.api [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141350, 'name': PowerOnVM_Task, 'duration_secs': 1.303242} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.307430] env[62692]: DEBUG oslo_vmware.api [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Task: {'id': task-1141349, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.871344} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.308110] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-208c48a1-0b06-4d48-bf0d-da956991ca93 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.312378] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 922.313184] env[62692]: INFO nova.compute.manager [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Took 5.92 seconds to spawn the instance on the hypervisor. [ 922.313381] env[62692]: DEBUG nova.compute.manager [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 922.314151] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 922.314151] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 922.314151] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 922.314358] env[62692]: INFO nova.compute.manager [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Took 2.24 seconds to destroy the instance on the hypervisor. [ 922.314646] env[62692]: DEBUG oslo.service.loopingcall [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.320593] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9927180f-aad6-4aa2-a1c1-6d1b57d2899a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.326373] env[62692]: DEBUG nova.compute.manager [-] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 922.326485] env[62692]: DEBUG nova.network.neutron [-] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 922.330474] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 922.330474] env[62692]: value = "task-1141352" [ 922.330474] env[62692]: _type = "Task" [ 922.330474] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.335205] env[62692]: DEBUG nova.network.neutron [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Successfully updated port: 7cf21363-a84c-4404-b293-ccee3ea80385 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 922.349843] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff799e46-4ed3-44b2-9a7f-80dd9ee9df69 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.359175] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 922.359175] env[62692]: value = "task-1141353" [ 922.359175] env[62692]: _type = "Task" [ 922.359175] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.380474] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141352, 'name': CreateVM_Task} progress is 15%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.386598] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141353, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.710868] env[62692]: DEBUG oslo_concurrency.lockutils [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.047s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.713620] env[62692]: DEBUG nova.compute.manager [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 922.716976] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.957s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.717413] env[62692]: DEBUG nova.objects.instance [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Lazy-loading 'resources' on Instance uuid b389b14d-6dcc-4b26-84f8-38b952bf6ecd {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 922.730239] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Acquiring lock "f44e2ee2-6321-4237-9644-a19b5e90ae16" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.732021] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Lock "f44e2ee2-6321-4237-9644-a19b5e90ae16" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.732021] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Acquiring lock "f44e2ee2-6321-4237-9644-a19b5e90ae16-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.732021] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Lock "f44e2ee2-6321-4237-9644-a19b5e90ae16-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.732591] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Lock "f44e2ee2-6321-4237-9644-a19b5e90ae16-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.740030] env[62692]: INFO nova.compute.manager [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Terminating instance [ 922.743757] env[62692]: DEBUG nova.compute.manager [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 922.744008] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 922.750675] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-35bb4feb-cc93-4650-871f-7b971f804239 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.765136] env[62692]: DEBUG oslo_concurrency.lockutils [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.779023] env[62692]: DEBUG oslo_vmware.api [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Waiting for the task: (returnval){ [ 922.779023] env[62692]: value = "task-1141355" [ 922.779023] env[62692]: _type = "Task" [ 922.779023] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.794864] env[62692]: DEBUG oslo_vmware.api [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Task: {'id': task-1141355, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.838317] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "refresh_cache-b23d3484-4fc5-46b2-8e8c-bdaab087f046" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.838317] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquired lock "refresh_cache-b23d3484-4fc5-46b2-8e8c-bdaab087f046" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.839031] env[62692]: DEBUG nova.network.neutron [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 922.848924] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141352, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.881368] env[62692]: INFO nova.compute.manager [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Took 42.20 seconds to build instance. [ 922.889374] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141353, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.962133] env[62692]: DEBUG nova.compute.manager [req-e1a0bfef-34cf-405c-afcf-8c2dab2f263f req-21000564-8c90-445a-85db-0b9336f67fa0 service nova] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Received event network-vif-deleted-b81f3e56-2a23-40b7-8153-e53be84131ca {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 922.962477] env[62692]: INFO nova.compute.manager [req-e1a0bfef-34cf-405c-afcf-8c2dab2f263f req-21000564-8c90-445a-85db-0b9336f67fa0 service nova] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Neutron deleted interface b81f3e56-2a23-40b7-8153-e53be84131ca; detaching it from the instance and deleting it from the info cache [ 922.962862] env[62692]: DEBUG nova.network.neutron [req-e1a0bfef-34cf-405c-afcf-8c2dab2f263f req-21000564-8c90-445a-85db-0b9336f67fa0 service nova] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Updating instance_info_cache with network_info: [{"id": "bfd79ef9-4a10-444c-96eb-1a4f4273a5d7", "address": "fa:16:3e:83:2a:f6", "network": {"id": "4c74b409-989a-4073-9def-fc9e2e3cc154", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1736671348", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.146", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fed583a514c4043a4b5772c73e21aff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31e77685-b4dd-4810-80ef-24115ea9ea62", "external-id": "nsx-vlan-transportzone-56", "segmentation_id": 56, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfd79ef9-4a", "ovs_interfaceid": "bfd79ef9-4a10-444c-96eb-1a4f4273a5d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.087629] env[62692]: DEBUG nova.network.neutron [req-e2caf092-e2e8-458e-9dcc-6ee4555adc5d req-90df103d-92e2-46b5-93d0-7605fb0be0b6 service nova] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Updated VIF entry in instance network info cache for port 0047b080-5217-4718-a134-a87f4556102b. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 923.087629] env[62692]: DEBUG nova.network.neutron [req-e2caf092-e2e8-458e-9dcc-6ee4555adc5d req-90df103d-92e2-46b5-93d0-7605fb0be0b6 service nova] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Updating instance_info_cache with network_info: [{"id": "0047b080-5217-4718-a134-a87f4556102b", "address": "fa:16:3e:0d:46:60", "network": {"id": "0e06e04a-5213-4d06-b3f1-b2b8f437e6d0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1533469056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "266c39fe585d41079d0278ecd1d1c6a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "678ebbe4-4c53-4eaf-a689-93981310f37d", "external-id": "nsx-vlan-transportzone-443", "segmentation_id": 443, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0047b080-52", "ovs_interfaceid": "0047b080-5217-4718-a134-a87f4556102b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.233846] env[62692]: DEBUG nova.compute.utils [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 923.235616] env[62692]: DEBUG nova.compute.manager [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 923.236491] env[62692]: DEBUG nova.network.neutron [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 923.265477] env[62692]: INFO nova.compute.manager [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Rebuilding instance [ 923.292269] env[62692]: DEBUG oslo_vmware.api [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Task: {'id': task-1141355, 'name': PowerOffVM_Task, 'duration_secs': 0.310284} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.292679] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 923.292679] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Volume detach. Driver type: vmdk {{(pid=62692) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 923.292826] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-248898', 'volume_id': '4e85e9f0-749c-4394-8d51-fe35fa42cd6d', 'name': 'volume-4e85e9f0-749c-4394-8d51-fe35fa42cd6d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f44e2ee2-6321-4237-9644-a19b5e90ae16', 'attached_at': '', 'detached_at': '', 'volume_id': '4e85e9f0-749c-4394-8d51-fe35fa42cd6d', 'serial': '4e85e9f0-749c-4394-8d51-fe35fa42cd6d'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 923.293596] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-919b57f1-3ddd-4e68-875b-c1a279d0e177 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.327964] env[62692]: DEBUG nova.policy [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '44a368c20ef24ade9a1191614b4a27da', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9016730e616045b4acd78ba6816f87a3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 923.330438] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46a94d5c-e36a-401b-905f-45ad4c36eebb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.336341] env[62692]: DEBUG nova.compute.manager [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 923.337706] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e652c25-1cef-4182-b18b-af6492bd4c97 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.359815] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141352, 'name': CreateVM_Task, 'duration_secs': 0.90558} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.360691] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71d4a2b0-8e66-4648-b659-6255566b48b8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.363416] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 923.366621] env[62692]: DEBUG oslo_concurrency.lockutils [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.366803] env[62692]: DEBUG oslo_concurrency.lockutils [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.367158] env[62692]: DEBUG oslo_concurrency.lockutils [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 923.368097] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6df55b9-2314-4250-b0f9-2a9c8af0f872 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.387367] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a25c0d0a-a563-4989-b783-708cb8e62807 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Lock "fac426b6-747a-455e-85a1-44e71b432180" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.945s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.393734] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03227009-27ce-45f5-ba24-f2c1578db2be {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.397504] env[62692]: DEBUG nova.network.neutron [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 923.404993] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141353, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.405187] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 923.405187] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52675182-1b48-a74e-338f-6a26f5bdec46" [ 923.405187] env[62692]: _type = "Task" [ 923.405187] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.422293] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] The volume has not been displaced from its original location: [datastore2] volume-4e85e9f0-749c-4394-8d51-fe35fa42cd6d/volume-4e85e9f0-749c-4394-8d51-fe35fa42cd6d.vmdk. No consolidation needed. {{(pid=62692) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 923.427743] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Reconfiguring VM instance instance-0000003d to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 923.433488] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-41e1d67e-10f6-4a65-9c05-1a48931b4486 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.450964] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52675182-1b48-a74e-338f-6a26f5bdec46, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.458952] env[62692]: DEBUG oslo_vmware.api [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Waiting for the task: (returnval){ [ 923.458952] env[62692]: value = "task-1141356" [ 923.458952] env[62692]: _type = "Task" [ 923.458952] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.465767] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-42755d6c-0b48-4d50-9948-5d5ff064dbc3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.474407] env[62692]: DEBUG oslo_vmware.api [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Task: {'id': task-1141356, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.482196] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb35a799-2663-4a8e-a3e2-ed4ee0e89a92 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.530013] env[62692]: DEBUG nova.compute.manager [req-e1a0bfef-34cf-405c-afcf-8c2dab2f263f req-21000564-8c90-445a-85db-0b9336f67fa0 service nova] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Detach interface failed, port_id=b81f3e56-2a23-40b7-8153-e53be84131ca, reason: Instance 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 923.590564] env[62692]: DEBUG oslo_concurrency.lockutils [req-e2caf092-e2e8-458e-9dcc-6ee4555adc5d req-90df103d-92e2-46b5-93d0-7605fb0be0b6 service nova] Releasing lock "refresh_cache-0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.660185] env[62692]: DEBUG nova.network.neutron [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Updating instance_info_cache with network_info: [{"id": "7cf21363-a84c-4404-b293-ccee3ea80385", "address": "fa:16:3e:ce:0c:7f", "network": {"id": "d0e00428-1d04-4b63-8975-7f3085784358", "bridge": "br-int", "label": "tempest-ServersTestJSON-408035146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e13f8888d5e84eed8247101558e29a56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7cf21363-a8", "ovs_interfaceid": "7cf21363-a84c-4404-b293-ccee3ea80385", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.743420] env[62692]: DEBUG nova.compute.manager [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 923.749197] env[62692]: DEBUG nova.network.neutron [-] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.760175] env[62692]: DEBUG nova.network.neutron [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Successfully created port: 1ff8b50f-db1f-4a27-8ce4-228ab7446c64 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 923.861340] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 923.862129] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3339532d-4444-4736-8031-68a89d703546 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.868451] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95cd0f97-8a3e-4f05-b7bf-7c6234bf8a27 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.876648] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 923.876648] env[62692]: value = "task-1141357" [ 923.876648] env[62692]: _type = "Task" [ 923.876648] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.886721] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141353, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.888284] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8572ef9-2bd8-4b41-9ec1-c97725d759d9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.895570] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141357, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.897202] env[62692]: DEBUG nova.compute.manager [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 923.930376] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3dd9d7f-25d4-46db-a812-978b2ee8b067 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.940231] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52675182-1b48-a74e-338f-6a26f5bdec46, 'name': SearchDatastore_Task, 'duration_secs': 0.095969} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.945047] env[62692]: DEBUG oslo_concurrency.lockutils [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.945047] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 923.945047] env[62692]: DEBUG oslo_concurrency.lockutils [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.945047] env[62692]: DEBUG oslo_concurrency.lockutils [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.945047] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 923.945487] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c9ac16a0-7954-46b3-98f7-3bd4ae2cc5f2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.948525] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08bb6f51-6a8c-4d1d-b294-62bd1aa5e5f9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.968018] env[62692]: DEBUG nova.compute.provider_tree [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 923.978159] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 923.978393] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 923.983361] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d77bef2f-2a63-44af-9835-75698f62eaec {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.986641] env[62692]: DEBUG oslo_vmware.api [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Task: {'id': task-1141356, 'name': ReconfigVM_Task, 'duration_secs': 0.207643} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.987415] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Reconfigured VM instance instance-0000003d to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 923.993645] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cadfd14d-f872-4c88-bb93-3085338cb0e6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.008130] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 924.008130] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]523f0b4c-72af-8346-45c9-af6491b03934" [ 924.008130] env[62692]: _type = "Task" [ 924.008130] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.018247] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]523f0b4c-72af-8346-45c9-af6491b03934, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.019777] env[62692]: DEBUG oslo_vmware.api [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Waiting for the task: (returnval){ [ 924.019777] env[62692]: value = "task-1141358" [ 924.019777] env[62692]: _type = "Task" [ 924.019777] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.027752] env[62692]: DEBUG nova.compute.manager [req-280593b6-b583-48b2-915c-7fac6d9e891b req-916102ae-d78c-4ae2-b250-d7b6ce04a39f service nova] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Received event network-vif-plugged-7cf21363-a84c-4404-b293-ccee3ea80385 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 924.027993] env[62692]: DEBUG oslo_concurrency.lockutils [req-280593b6-b583-48b2-915c-7fac6d9e891b req-916102ae-d78c-4ae2-b250-d7b6ce04a39f service nova] Acquiring lock "b23d3484-4fc5-46b2-8e8c-bdaab087f046-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.028229] env[62692]: DEBUG oslo_concurrency.lockutils [req-280593b6-b583-48b2-915c-7fac6d9e891b req-916102ae-d78c-4ae2-b250-d7b6ce04a39f service nova] Lock "b23d3484-4fc5-46b2-8e8c-bdaab087f046-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.028405] env[62692]: DEBUG oslo_concurrency.lockutils [req-280593b6-b583-48b2-915c-7fac6d9e891b req-916102ae-d78c-4ae2-b250-d7b6ce04a39f service nova] Lock "b23d3484-4fc5-46b2-8e8c-bdaab087f046-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.028577] env[62692]: DEBUG nova.compute.manager [req-280593b6-b583-48b2-915c-7fac6d9e891b req-916102ae-d78c-4ae2-b250-d7b6ce04a39f service nova] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] No waiting events found dispatching network-vif-plugged-7cf21363-a84c-4404-b293-ccee3ea80385 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 924.028746] env[62692]: WARNING nova.compute.manager [req-280593b6-b583-48b2-915c-7fac6d9e891b req-916102ae-d78c-4ae2-b250-d7b6ce04a39f service nova] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Received unexpected event network-vif-plugged-7cf21363-a84c-4404-b293-ccee3ea80385 for instance with vm_state building and task_state spawning. [ 924.028992] env[62692]: DEBUG nova.compute.manager [req-280593b6-b583-48b2-915c-7fac6d9e891b req-916102ae-d78c-4ae2-b250-d7b6ce04a39f service nova] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Received event network-changed-7cf21363-a84c-4404-b293-ccee3ea80385 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 924.029361] env[62692]: DEBUG nova.compute.manager [req-280593b6-b583-48b2-915c-7fac6d9e891b req-916102ae-d78c-4ae2-b250-d7b6ce04a39f service nova] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Refreshing instance network info cache due to event network-changed-7cf21363-a84c-4404-b293-ccee3ea80385. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 924.029613] env[62692]: DEBUG oslo_concurrency.lockutils [req-280593b6-b583-48b2-915c-7fac6d9e891b req-916102ae-d78c-4ae2-b250-d7b6ce04a39f service nova] Acquiring lock "refresh_cache-b23d3484-4fc5-46b2-8e8c-bdaab087f046" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.034247] env[62692]: DEBUG oslo_vmware.api [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Task: {'id': task-1141358, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.167053] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Releasing lock "refresh_cache-b23d3484-4fc5-46b2-8e8c-bdaab087f046" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.167568] env[62692]: DEBUG nova.compute.manager [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Instance network_info: |[{"id": "7cf21363-a84c-4404-b293-ccee3ea80385", "address": "fa:16:3e:ce:0c:7f", "network": {"id": "d0e00428-1d04-4b63-8975-7f3085784358", "bridge": "br-int", "label": "tempest-ServersTestJSON-408035146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e13f8888d5e84eed8247101558e29a56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7cf21363-a8", "ovs_interfaceid": "7cf21363-a84c-4404-b293-ccee3ea80385", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 924.168046] env[62692]: DEBUG oslo_concurrency.lockutils [req-280593b6-b583-48b2-915c-7fac6d9e891b req-916102ae-d78c-4ae2-b250-d7b6ce04a39f service nova] Acquired lock "refresh_cache-b23d3484-4fc5-46b2-8e8c-bdaab087f046" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.168367] env[62692]: DEBUG nova.network.neutron [req-280593b6-b583-48b2-915c-7fac6d9e891b req-916102ae-d78c-4ae2-b250-d7b6ce04a39f service nova] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Refreshing network info cache for port 7cf21363-a84c-4404-b293-ccee3ea80385 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 924.170284] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ce:0c:7f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7cf21363-a84c-4404-b293-ccee3ea80385', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 924.184388] env[62692]: DEBUG oslo.service.loopingcall [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 924.188334] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 924.189183] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-67f83dae-d722-46b9-9f61-d05598a7c214 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.221163] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 924.221163] env[62692]: value = "task-1141359" [ 924.221163] env[62692]: _type = "Task" [ 924.221163] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.234220] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141359, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.250415] env[62692]: INFO nova.compute.manager [-] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Took 1.92 seconds to deallocate network for instance. [ 924.387038] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141353, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.399419] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141357, 'name': PowerOffVM_Task, 'duration_secs': 0.14045} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.399831] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 924.400077] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 924.401340] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26bf9ae5-e74f-4578-947f-e34cc337f032 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.414837] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 924.415183] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-78ee9cb9-b51b-475e-b289-318206aed029 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.429508] env[62692]: DEBUG oslo_concurrency.lockutils [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.448979] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 924.449381] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 924.449587] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Deleting the datastore file [datastore1] fac426b6-747a-455e-85a1-44e71b432180 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 924.449948] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fc5fa5f1-bbbd-4a4e-a897-fb3dbc59923b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.459703] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 924.459703] env[62692]: value = "task-1141361" [ 924.459703] env[62692]: _type = "Task" [ 924.459703] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.474935] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141361, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.494089] env[62692]: ERROR nova.scheduler.client.report [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] [req-04021359-08d9-41bf-af43-d48c6a1c8df5] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 470758a8-5efa-4875-aad5-f512a727752f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-04021359-08d9-41bf-af43-d48c6a1c8df5"}]} [ 924.504502] env[62692]: DEBUG nova.network.neutron [req-280593b6-b583-48b2-915c-7fac6d9e891b req-916102ae-d78c-4ae2-b250-d7b6ce04a39f service nova] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Updated VIF entry in instance network info cache for port 7cf21363-a84c-4404-b293-ccee3ea80385. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 924.504889] env[62692]: DEBUG nova.network.neutron [req-280593b6-b583-48b2-915c-7fac6d9e891b req-916102ae-d78c-4ae2-b250-d7b6ce04a39f service nova] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Updating instance_info_cache with network_info: [{"id": "7cf21363-a84c-4404-b293-ccee3ea80385", "address": "fa:16:3e:ce:0c:7f", "network": {"id": "d0e00428-1d04-4b63-8975-7f3085784358", "bridge": "br-int", "label": "tempest-ServersTestJSON-408035146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e13f8888d5e84eed8247101558e29a56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7cf21363-a8", "ovs_interfaceid": "7cf21363-a84c-4404-b293-ccee3ea80385", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.514187] env[62692]: DEBUG nova.scheduler.client.report [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Refreshing inventories for resource provider 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 924.524254] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]523f0b4c-72af-8346-45c9-af6491b03934, 'name': SearchDatastore_Task, 'duration_secs': 0.098784} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.528781] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30582829-00a8-4a4a-ba76-965200bf7efe {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.534125] env[62692]: DEBUG nova.scheduler.client.report [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Updating ProviderTree inventory for provider 470758a8-5efa-4875-aad5-f512a727752f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 924.534459] env[62692]: DEBUG nova.compute.provider_tree [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 924.542216] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 924.542216] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]524d8a09-ebf1-a86a-5094-9d00ad5aaa23" [ 924.542216] env[62692]: _type = "Task" [ 924.542216] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.542455] env[62692]: DEBUG oslo_vmware.api [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Task: {'id': task-1141358, 'name': ReconfigVM_Task, 'duration_secs': 0.207032} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.542810] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-248898', 'volume_id': '4e85e9f0-749c-4394-8d51-fe35fa42cd6d', 'name': 'volume-4e85e9f0-749c-4394-8d51-fe35fa42cd6d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f44e2ee2-6321-4237-9644-a19b5e90ae16', 'attached_at': '', 'detached_at': '', 'volume_id': '4e85e9f0-749c-4394-8d51-fe35fa42cd6d', 'serial': '4e85e9f0-749c-4394-8d51-fe35fa42cd6d'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 924.543115] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 924.546856] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ed5d09d-8d51-4b7b-ae1c-dcc04fbfc1da {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.553832] env[62692]: DEBUG nova.scheduler.client.report [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Refreshing aggregate associations for resource provider 470758a8-5efa-4875-aad5-f512a727752f, aggregates: None {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 924.561866] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]524d8a09-ebf1-a86a-5094-9d00ad5aaa23, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.562230] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 924.562505] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a2ec3158-59a1-4bc5-93e8-2000dca6bd0a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.580744] env[62692]: DEBUG nova.scheduler.client.report [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Refreshing trait associations for resource provider 470758a8-5efa-4875-aad5-f512a727752f, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 924.640397] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 924.641722] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 924.642189] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Deleting the datastore file [datastore2] f44e2ee2-6321-4237-9644-a19b5e90ae16 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 924.642639] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f106b5e1-2096-4433-9d4e-4e1dfb836c40 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.654919] env[62692]: DEBUG oslo_vmware.api [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Waiting for the task: (returnval){ [ 924.654919] env[62692]: value = "task-1141363" [ 924.654919] env[62692]: _type = "Task" [ 924.654919] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.668781] env[62692]: DEBUG oslo_vmware.api [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Task: {'id': task-1141363, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.736811] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141359, 'name': CreateVM_Task, 'duration_secs': 0.517615} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.739878] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 924.744017] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.744017] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.744017] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 924.744017] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4391476-3eec-4ca7-bfca-3fbbe29c6e22 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.747842] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 924.747842] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]528df57e-d3b3-d3cb-6f2d-015cc448ac4a" [ 924.747842] env[62692]: _type = "Task" [ 924.747842] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.756612] env[62692]: DEBUG nova.compute.manager [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 924.766317] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.766317] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]528df57e-d3b3-d3cb-6f2d-015cc448ac4a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.793411] env[62692]: DEBUG nova.virt.hardware [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 924.794483] env[62692]: DEBUG nova.virt.hardware [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 924.794483] env[62692]: DEBUG nova.virt.hardware [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 924.794483] env[62692]: DEBUG nova.virt.hardware [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 924.794483] env[62692]: DEBUG nova.virt.hardware [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 924.794483] env[62692]: DEBUG nova.virt.hardware [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 924.794707] env[62692]: DEBUG nova.virt.hardware [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 924.794707] env[62692]: DEBUG nova.virt.hardware [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 924.794861] env[62692]: DEBUG nova.virt.hardware [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 924.795039] env[62692]: DEBUG nova.virt.hardware [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 924.795417] env[62692]: DEBUG nova.virt.hardware [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 924.796328] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8cc769e-5325-4691-8a3c-155e29cc52c2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.809377] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d730e28-1ac1-4bf0-97e6-c77803836b05 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.879744] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141353, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.971938] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141361, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.008262] env[62692]: DEBUG oslo_concurrency.lockutils [req-280593b6-b583-48b2-915c-7fac6d9e891b req-916102ae-d78c-4ae2-b250-d7b6ce04a39f service nova] Releasing lock "refresh_cache-b23d3484-4fc5-46b2-8e8c-bdaab087f046" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.053121] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]524d8a09-ebf1-a86a-5094-9d00ad5aaa23, 'name': SearchDatastore_Task, 'duration_secs': 0.096672} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.055415] env[62692]: DEBUG oslo_concurrency.lockutils [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.055685] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d/0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 925.056190] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-03d28512-dd3f-4a50-8207-228f0af0c97a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.063593] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 925.063593] env[62692]: value = "task-1141364" [ 925.063593] env[62692]: _type = "Task" [ 925.063593] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.074241] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141364, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.079441] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fab82c9-0c0f-46a5-bafa-e740a277703b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.086590] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e558ace8-23f5-47ef-addd-b3f051010262 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.094932] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea85a1e4-edbb-4f66-9d9a-1923c1786ac2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Volume attach. Driver type: vmdk {{(pid=62692) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 925.095226] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea85a1e4-edbb-4f66-9d9a-1923c1786ac2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-248983', 'volume_id': '14bc011a-54f2-4452-95c6-b9a4f2020742', 'name': 'volume-14bc011a-54f2-4452-95c6-b9a4f2020742', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3fd2959b-80cb-470c-8c2a-40b7630458ed', 'attached_at': '', 'detached_at': '', 'volume_id': '14bc011a-54f2-4452-95c6-b9a4f2020742', 'serial': '14bc011a-54f2-4452-95c6-b9a4f2020742'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 925.095991] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d76a18f-6ea7-40f3-a7eb-024b824301ac {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.124267] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cb3948e-d483-45da-9a25-c6e9657681e7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.138942] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a32482e2-2567-45b0-920c-edb83fb0e433 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.147197] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d386448f-727e-46d6-9cb9-1405d83cf1cc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.184779] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea85a1e4-edbb-4f66-9d9a-1923c1786ac2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] volume-14bc011a-54f2-4452-95c6-b9a4f2020742/volume-14bc011a-54f2-4452-95c6-b9a4f2020742.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 925.189934] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-faf812ef-de72-4667-9ee3-1d128f352f04 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.216439] env[62692]: DEBUG nova.compute.manager [req-d318a88f-e31e-4514-b37c-d8b76302cbfb req-49477404-7b07-45f3-ae16-5bf6f97b8955 service nova] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Received event network-vif-deleted-bfd79ef9-4a10-444c-96eb-1a4f4273a5d7 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 925.216865] env[62692]: DEBUG nova.compute.provider_tree [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 925.224063] env[62692]: DEBUG oslo_vmware.api [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Task: {'id': task-1141363, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.458549} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.225637] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 925.225637] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 925.225799] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 925.227020] env[62692]: INFO nova.compute.manager [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Took 2.48 seconds to destroy the instance on the hypervisor. [ 925.227020] env[62692]: DEBUG oslo.service.loopingcall [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 925.227020] env[62692]: DEBUG oslo_vmware.api [None req-ea85a1e4-edbb-4f66-9d9a-1923c1786ac2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 925.227020] env[62692]: value = "task-1141365" [ 925.227020] env[62692]: _type = "Task" [ 925.227020] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.227020] env[62692]: DEBUG nova.compute.manager [-] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 925.227020] env[62692]: DEBUG nova.network.neutron [-] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 925.241078] env[62692]: DEBUG oslo_vmware.api [None req-ea85a1e4-edbb-4f66-9d9a-1923c1786ac2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141365, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.259569] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]528df57e-d3b3-d3cb-6f2d-015cc448ac4a, 'name': SearchDatastore_Task, 'duration_secs': 0.026} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.259885] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.260136] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 925.260370] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.260511] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.260707] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 925.260968] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e0cf1d4b-0aac-4d9e-9924-ba923360783c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.272806] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 925.273325] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 925.273712] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08ec1b3a-b0d8-4bbe-aa85-f00a3c0d4da1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.280081] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 925.280081] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]525d4b31-6437-d48e-6679-b84a70ea09a1" [ 925.280081] env[62692]: _type = "Task" [ 925.280081] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.288852] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]525d4b31-6437-d48e-6679-b84a70ea09a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.363336] env[62692]: DEBUG nova.network.neutron [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Successfully updated port: 1ff8b50f-db1f-4a27-8ce4-228ab7446c64 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 925.381150] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141353, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.558161} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.381546] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/427c2d72-be26-42e4-9f8c-ca628806da41/427c2d72-be26-42e4-9f8c-ca628806da41.vmdk to [datastore1] 44699df6-5262-4700-a096-8c0e450323d2/44699df6-5262-4700-a096-8c0e450323d2.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 925.382605] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-729a44e0-0906-4d1c-bdfa-2243730b78ac {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.407782] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 44699df6-5262-4700-a096-8c0e450323d2/44699df6-5262-4700-a096-8c0e450323d2.vmdk or device None with type streamOptimized {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 925.408104] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d7c7d7fb-3ea9-48a5-897e-c6e447bf7adc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.429476] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 925.429476] env[62692]: value = "task-1141366" [ 925.429476] env[62692]: _type = "Task" [ 925.429476] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.438380] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141366, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.471407] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141361, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.86928} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.472371] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 925.473015] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 925.473330] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 925.579982] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141364, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.739065] env[62692]: DEBUG oslo_vmware.api [None req-ea85a1e4-edbb-4f66-9d9a-1923c1786ac2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141365, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.773438] env[62692]: DEBUG nova.scheduler.client.report [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Updated inventory for provider 470758a8-5efa-4875-aad5-f512a727752f with generation 97 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 925.773735] env[62692]: DEBUG nova.compute.provider_tree [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Updating resource provider 470758a8-5efa-4875-aad5-f512a727752f generation from 97 to 98 during operation: update_inventory {{(pid=62692) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 925.773950] env[62692]: DEBUG nova.compute.provider_tree [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 925.793585] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]525d4b31-6437-d48e-6679-b84a70ea09a1, 'name': SearchDatastore_Task, 'duration_secs': 0.042891} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.794898] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a41dde1d-ef50-4db9-ba5e-fffb2e9dd8ff {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.802190] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 925.802190] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52e3bc6a-feb4-78a9-647a-d640e81587e6" [ 925.802190] env[62692]: _type = "Task" [ 925.802190] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.811133] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52e3bc6a-feb4-78a9-647a-d640e81587e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.865122] env[62692]: DEBUG oslo_concurrency.lockutils [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Acquiring lock "refresh_cache-8a56e6df-eea1-41f4-9360-4f06d2f516a3" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.865274] env[62692]: DEBUG oslo_concurrency.lockutils [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Acquired lock "refresh_cache-8a56e6df-eea1-41f4-9360-4f06d2f516a3" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.865385] env[62692]: DEBUG nova.network.neutron [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 925.943380] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141366, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.078011] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141364, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.129945] env[62692]: DEBUG nova.compute.manager [req-219b20d6-8072-4ea6-b90e-83deb0c1de3c req-40507192-96eb-4157-9fa9-12a7669ba205 service nova] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Received event network-vif-plugged-1ff8b50f-db1f-4a27-8ce4-228ab7446c64 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 926.129945] env[62692]: DEBUG oslo_concurrency.lockutils [req-219b20d6-8072-4ea6-b90e-83deb0c1de3c req-40507192-96eb-4157-9fa9-12a7669ba205 service nova] Acquiring lock "8a56e6df-eea1-41f4-9360-4f06d2f516a3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.130811] env[62692]: DEBUG oslo_concurrency.lockutils [req-219b20d6-8072-4ea6-b90e-83deb0c1de3c req-40507192-96eb-4157-9fa9-12a7669ba205 service nova] Lock "8a56e6df-eea1-41f4-9360-4f06d2f516a3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.132899] env[62692]: DEBUG oslo_concurrency.lockutils [req-219b20d6-8072-4ea6-b90e-83deb0c1de3c req-40507192-96eb-4157-9fa9-12a7669ba205 service nova] Lock "8a56e6df-eea1-41f4-9360-4f06d2f516a3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.133206] env[62692]: DEBUG nova.compute.manager [req-219b20d6-8072-4ea6-b90e-83deb0c1de3c req-40507192-96eb-4157-9fa9-12a7669ba205 service nova] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] No waiting events found dispatching network-vif-plugged-1ff8b50f-db1f-4a27-8ce4-228ab7446c64 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 926.133438] env[62692]: WARNING nova.compute.manager [req-219b20d6-8072-4ea6-b90e-83deb0c1de3c req-40507192-96eb-4157-9fa9-12a7669ba205 service nova] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Received unexpected event network-vif-plugged-1ff8b50f-db1f-4a27-8ce4-228ab7446c64 for instance with vm_state building and task_state spawning. [ 926.133630] env[62692]: DEBUG nova.compute.manager [req-219b20d6-8072-4ea6-b90e-83deb0c1de3c req-40507192-96eb-4157-9fa9-12a7669ba205 service nova] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Received event network-changed-1ff8b50f-db1f-4a27-8ce4-228ab7446c64 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 926.133793] env[62692]: DEBUG nova.compute.manager [req-219b20d6-8072-4ea6-b90e-83deb0c1de3c req-40507192-96eb-4157-9fa9-12a7669ba205 service nova] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Refreshing instance network info cache due to event network-changed-1ff8b50f-db1f-4a27-8ce4-228ab7446c64. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 926.133979] env[62692]: DEBUG oslo_concurrency.lockutils [req-219b20d6-8072-4ea6-b90e-83deb0c1de3c req-40507192-96eb-4157-9fa9-12a7669ba205 service nova] Acquiring lock "refresh_cache-8a56e6df-eea1-41f4-9360-4f06d2f516a3" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.240669] env[62692]: DEBUG oslo_vmware.api [None req-ea85a1e4-edbb-4f66-9d9a-1923c1786ac2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141365, 'name': ReconfigVM_Task, 'duration_secs': 0.812455} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.240980] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea85a1e4-edbb-4f66-9d9a-1923c1786ac2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Reconfigured VM instance instance-00000043 to attach disk [datastore1] volume-14bc011a-54f2-4452-95c6-b9a4f2020742/volume-14bc011a-54f2-4452-95c6-b9a4f2020742.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 926.246106] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe4c4fa9-14eb-4b19-a8b9-717de8a24c25 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.262605] env[62692]: DEBUG oslo_vmware.api [None req-ea85a1e4-edbb-4f66-9d9a-1923c1786ac2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 926.262605] env[62692]: value = "task-1141367" [ 926.262605] env[62692]: _type = "Task" [ 926.262605] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.271932] env[62692]: DEBUG oslo_vmware.api [None req-ea85a1e4-edbb-4f66-9d9a-1923c1786ac2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141367, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.279957] env[62692]: DEBUG nova.network.neutron [-] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.281377] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.565s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.283922] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 27.662s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.284130] env[62692]: DEBUG nova.objects.instance [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62692) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 926.307939] env[62692]: INFO nova.scheduler.client.report [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Deleted allocations for instance b389b14d-6dcc-4b26-84f8-38b952bf6ecd [ 926.318268] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52e3bc6a-feb4-78a9-647a-d640e81587e6, 'name': SearchDatastore_Task, 'duration_secs': 0.048171} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.318850] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.321034] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] b23d3484-4fc5-46b2-8e8c-bdaab087f046/b23d3484-4fc5-46b2-8e8c-bdaab087f046.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 926.321034] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f144d95c-4ba4-4fa2-a4ef-251bd670a88c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.328382] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 926.328382] env[62692]: value = "task-1141368" [ 926.328382] env[62692]: _type = "Task" [ 926.328382] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.338623] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141368, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.400566] env[62692]: DEBUG nova.network.neutron [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 926.441434] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141366, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.512824] env[62692]: DEBUG nova.virt.hardware [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 926.513104] env[62692]: DEBUG nova.virt.hardware [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 926.513273] env[62692]: DEBUG nova.virt.hardware [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 926.513465] env[62692]: DEBUG nova.virt.hardware [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 926.513618] env[62692]: DEBUG nova.virt.hardware [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 926.513770] env[62692]: DEBUG nova.virt.hardware [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 926.513984] env[62692]: DEBUG nova.virt.hardware [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 926.514164] env[62692]: DEBUG nova.virt.hardware [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 926.514337] env[62692]: DEBUG nova.virt.hardware [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 926.514527] env[62692]: DEBUG nova.virt.hardware [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 926.514724] env[62692]: DEBUG nova.virt.hardware [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 926.515756] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46ce9702-b924-4aef-9bb6-69ce9714ca4d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.524640] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36c81a29-58c9-4b8a-bbb4-d8d9feda5597 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.540849] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Instance VIF info [] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 926.546475] env[62692]: DEBUG oslo.service.loopingcall [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 926.547531] env[62692]: DEBUG nova.network.neutron [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Updating instance_info_cache with network_info: [{"id": "1ff8b50f-db1f-4a27-8ce4-228ab7446c64", "address": "fa:16:3e:84:0c:ad", "network": {"id": "3947088a-cf2e-47a8-89dd-89b92423ca70", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-198954239-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9016730e616045b4acd78ba6816f87a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f267bcdd-0daa-4337-9709-5fc060c267d8", "external-id": "nsx-vlan-transportzone-308", "segmentation_id": 308, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1ff8b50f-db", "ovs_interfaceid": "1ff8b50f-db1f-4a27-8ce4-228ab7446c64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.548872] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fac426b6-747a-455e-85a1-44e71b432180] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 926.549510] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-da6153ac-a9cb-4134-b60f-687da5f677f5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.572924] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 926.572924] env[62692]: value = "task-1141369" [ 926.572924] env[62692]: _type = "Task" [ 926.572924] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.575568] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141364, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.210197} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.578384] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d/0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 926.578605] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 926.579076] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-db726138-98c8-4376-b727-9aed661308c6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.586672] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141369, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.588267] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 926.588267] env[62692]: value = "task-1141370" [ 926.588267] env[62692]: _type = "Task" [ 926.588267] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.596803] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141370, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.773446] env[62692]: DEBUG oslo_vmware.api [None req-ea85a1e4-edbb-4f66-9d9a-1923c1786ac2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141367, 'name': ReconfigVM_Task, 'duration_secs': 0.159136} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.773771] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea85a1e4-edbb-4f66-9d9a-1923c1786ac2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-248983', 'volume_id': '14bc011a-54f2-4452-95c6-b9a4f2020742', 'name': 'volume-14bc011a-54f2-4452-95c6-b9a4f2020742', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3fd2959b-80cb-470c-8c2a-40b7630458ed', 'attached_at': '', 'detached_at': '', 'volume_id': '14bc011a-54f2-4452-95c6-b9a4f2020742', 'serial': '14bc011a-54f2-4452-95c6-b9a4f2020742'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 926.785100] env[62692]: INFO nova.compute.manager [-] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Took 1.56 seconds to deallocate network for instance. [ 926.823174] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e4b1f19a-9039-47e3-85d6-8ba1bba515ab tempest-ImagesNegativeTestJSON-1341147886 tempest-ImagesNegativeTestJSON-1341147886-project-member] Lock "b389b14d-6dcc-4b26-84f8-38b952bf6ecd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.518s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.840449] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141368, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.943394] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141366, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.050184] env[62692]: DEBUG oslo_concurrency.lockutils [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Releasing lock "refresh_cache-8a56e6df-eea1-41f4-9360-4f06d2f516a3" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.050561] env[62692]: DEBUG nova.compute.manager [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Instance network_info: |[{"id": "1ff8b50f-db1f-4a27-8ce4-228ab7446c64", "address": "fa:16:3e:84:0c:ad", "network": {"id": "3947088a-cf2e-47a8-89dd-89b92423ca70", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-198954239-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9016730e616045b4acd78ba6816f87a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f267bcdd-0daa-4337-9709-5fc060c267d8", "external-id": "nsx-vlan-transportzone-308", "segmentation_id": 308, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1ff8b50f-db", "ovs_interfaceid": "1ff8b50f-db1f-4a27-8ce4-228ab7446c64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 927.050888] env[62692]: DEBUG oslo_concurrency.lockutils [req-219b20d6-8072-4ea6-b90e-83deb0c1de3c req-40507192-96eb-4157-9fa9-12a7669ba205 service nova] Acquired lock "refresh_cache-8a56e6df-eea1-41f4-9360-4f06d2f516a3" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.051098] env[62692]: DEBUG nova.network.neutron [req-219b20d6-8072-4ea6-b90e-83deb0c1de3c req-40507192-96eb-4157-9fa9-12a7669ba205 service nova] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Refreshing network info cache for port 1ff8b50f-db1f-4a27-8ce4-228ab7446c64 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 927.052398] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:84:0c:ad', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f267bcdd-0daa-4337-9709-5fc060c267d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1ff8b50f-db1f-4a27-8ce4-228ab7446c64', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 927.061065] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Creating folder: Project (9016730e616045b4acd78ba6816f87a3). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 927.065249] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6d1ab970-c1e7-4f66-a830-81546c90465d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.080721] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Created folder: Project (9016730e616045b4acd78ba6816f87a3) in parent group-v248868. [ 927.081231] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Creating folder: Instances. Parent ref: group-v248987. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 927.081626] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-58a1a3c6-70fe-41f4-86da-36f29b1570bb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.086794] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141369, 'name': CreateVM_Task, 'duration_secs': 0.300982} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.086964] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fac426b6-747a-455e-85a1-44e71b432180] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 927.087777] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.087777] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.087955] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 927.088502] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4fe75a9d-444f-4468-86ba-fa2d5564e644 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.093921] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Created folder: Instances in parent group-v248987. [ 927.094174] env[62692]: DEBUG oslo.service.loopingcall [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 927.094669] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 927.094882] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-40155985-8239-446d-8701-c9093cc1e258 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.115365] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 927.115365] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5288a756-869b-ce33-911b-598793f73b9e" [ 927.115365] env[62692]: _type = "Task" [ 927.115365] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.116293] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141370, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070899} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.116293] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 927.120588] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6fdf60b-9adb-4a15-8127-1c5b0ef6f209 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.123963] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 927.123963] env[62692]: value = "task-1141373" [ 927.123963] env[62692]: _type = "Task" [ 927.123963] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.149275] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d/0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 927.149604] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5288a756-869b-ce33-911b-598793f73b9e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.150615] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9cdb8397-0cc3-4b5b-8bf9-838fc79577f0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.170236] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141373, 'name': CreateVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.176549] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 927.176549] env[62692]: value = "task-1141374" [ 927.176549] env[62692]: _type = "Task" [ 927.176549] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.184867] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141374, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.219039] env[62692]: DEBUG nova.compute.manager [req-1673d07d-6f62-499d-a217-d383307e7c65 req-c7f47731-ffd0-4819-8a1c-0cd42831821d service nova] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Received event network-vif-deleted-49431714-cc36-41a1-a843-809653e38ccc {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 927.293047] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ca45a4a0-af59-4557-af8d-9327eaab0665 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.294238] env[62692]: DEBUG oslo_concurrency.lockutils [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.533s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.294400] env[62692]: DEBUG nova.objects.instance [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lazy-loading 'resources' on Instance uuid 433f740c-bde4-4d33-baca-5a7ac8179da1 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.331020] env[62692]: DEBUG nova.network.neutron [req-219b20d6-8072-4ea6-b90e-83deb0c1de3c req-40507192-96eb-4157-9fa9-12a7669ba205 service nova] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Updated VIF entry in instance network info cache for port 1ff8b50f-db1f-4a27-8ce4-228ab7446c64. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 927.331502] env[62692]: DEBUG nova.network.neutron [req-219b20d6-8072-4ea6-b90e-83deb0c1de3c req-40507192-96eb-4157-9fa9-12a7669ba205 service nova] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Updating instance_info_cache with network_info: [{"id": "1ff8b50f-db1f-4a27-8ce4-228ab7446c64", "address": "fa:16:3e:84:0c:ad", "network": {"id": "3947088a-cf2e-47a8-89dd-89b92423ca70", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-198954239-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9016730e616045b4acd78ba6816f87a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f267bcdd-0daa-4337-9709-5fc060c267d8", "external-id": "nsx-vlan-transportzone-308", "segmentation_id": 308, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1ff8b50f-db", "ovs_interfaceid": "1ff8b50f-db1f-4a27-8ce4-228ab7446c64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.346942] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141368, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.352883] env[62692]: INFO nova.compute.manager [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Took 0.57 seconds to detach 1 volumes for instance. [ 927.355031] env[62692]: DEBUG nova.compute.manager [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Deleting volume: 4e85e9f0-749c-4394-8d51-fe35fa42cd6d {{(pid=62692) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 927.444507] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141366, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.629852] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5288a756-869b-ce33-911b-598793f73b9e, 'name': SearchDatastore_Task, 'duration_secs': 0.040311} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.634439] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.634849] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 927.635430] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.635736] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.636111] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 927.636525] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f13f8afc-0876-4608-9208-2c520781da4c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.649153] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141373, 'name': CreateVM_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.655338] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 927.656542] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 927.657777] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-561a9dae-d34b-4fd1-9576-4eba28248bff {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.666593] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 927.666593] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52251bd0-2872-56da-e1e1-ae1f4ee52bfe" [ 927.666593] env[62692]: _type = "Task" [ 927.666593] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.677208] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52251bd0-2872-56da-e1e1-ae1f4ee52bfe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.689996] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141374, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.828540] env[62692]: DEBUG nova.objects.instance [None req-ea85a1e4-edbb-4f66-9d9a-1923c1786ac2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lazy-loading 'flavor' on Instance uuid 3fd2959b-80cb-470c-8c2a-40b7630458ed {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.840418] env[62692]: DEBUG oslo_concurrency.lockutils [req-219b20d6-8072-4ea6-b90e-83deb0c1de3c req-40507192-96eb-4157-9fa9-12a7669ba205 service nova] Releasing lock "refresh_cache-8a56e6df-eea1-41f4-9360-4f06d2f516a3" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.851021] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141368, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.896609] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.952762] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141366, 'name': ReconfigVM_Task, 'duration_secs': 2.07563} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.952762] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 44699df6-5262-4700-a096-8c0e450323d2/44699df6-5262-4700-a096-8c0e450323d2.vmdk or device None with type streamOptimized {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 927.952762] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-59eef6ec-fb68-424e-8107-7bcf5360cb4b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.962543] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 927.962543] env[62692]: value = "task-1141376" [ 927.962543] env[62692]: _type = "Task" [ 927.962543] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.977313] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141376, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.142848] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141373, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.178903] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52251bd0-2872-56da-e1e1-ae1f4ee52bfe, 'name': SearchDatastore_Task, 'duration_secs': 0.064505} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.182661] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a2525d2-2bd6-47da-b930-c25673da30b5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.195549] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141374, 'name': ReconfigVM_Task, 'duration_secs': 0.891144} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.196125] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Reconfigured VM instance instance-00000047 to attach disk [datastore1] 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d/0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 928.196632] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 928.196632] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]521f4366-77a6-ee61-89f2-7d1d41de4df8" [ 928.196632] env[62692]: _type = "Task" [ 928.196632] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.196869] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-139a81fc-7ba1-4a02-aac4-70a770524a4e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.210666] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]521f4366-77a6-ee61-89f2-7d1d41de4df8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.212656] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 928.212656] env[62692]: value = "task-1141377" [ 928.212656] env[62692]: _type = "Task" [ 928.212656] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.229622] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141377, 'name': Rename_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.292922] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cf57de8-d09a-4257-8834-1f8c6b356a28 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.301653] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-661998fc-9877-4cdb-a984-a94f5277255a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.343932] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ea85a1e4-edbb-4f66-9d9a-1923c1786ac2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "3fd2959b-80cb-470c-8c2a-40b7630458ed" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.383s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.349549] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eff1cbc-5f62-4bb6-a095-b90b152d9e1a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.363222] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f52af7b8-8911-457c-8068-7a25454f28c6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.367686] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141368, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.664447} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.367954] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] b23d3484-4fc5-46b2-8e8c-bdaab087f046/b23d3484-4fc5-46b2-8e8c-bdaab087f046.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 928.368192] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 928.368783] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7758af95-1ba3-4e21-8c0c-55456ff8ac4a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.384738] env[62692]: DEBUG nova.compute.provider_tree [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 928.387448] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 928.387448] env[62692]: value = "task-1141378" [ 928.387448] env[62692]: _type = "Task" [ 928.387448] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.397717] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141378, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.473197] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141376, 'name': Rename_Task, 'duration_secs': 0.384149} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.473482] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 928.473730] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dec02afb-80c1-49f9-b72f-1bc6451f8aab {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.481628] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 928.481628] env[62692]: value = "task-1141379" [ 928.481628] env[62692]: _type = "Task" [ 928.481628] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.490881] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141379, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.585449] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "3fd2959b-80cb-470c-8c2a-40b7630458ed" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.585834] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "3fd2959b-80cb-470c-8c2a-40b7630458ed" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.586067] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "3fd2959b-80cb-470c-8c2a-40b7630458ed-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.586365] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "3fd2959b-80cb-470c-8c2a-40b7630458ed-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.586581] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "3fd2959b-80cb-470c-8c2a-40b7630458ed-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.588765] env[62692]: INFO nova.compute.manager [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Terminating instance [ 928.590833] env[62692]: DEBUG nova.compute.manager [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 928.591066] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 928.591312] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-de2fd58b-7b01-455a-8e23-da97876df61f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.600053] env[62692]: DEBUG oslo_vmware.api [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 928.600053] env[62692]: value = "task-1141380" [ 928.600053] env[62692]: _type = "Task" [ 928.600053] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.610140] env[62692]: DEBUG oslo_vmware.api [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141380, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.644431] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141373, 'name': CreateVM_Task, 'duration_secs': 1.065524} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.644703] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 928.645654] env[62692]: DEBUG oslo_concurrency.lockutils [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.645756] env[62692]: DEBUG oslo_concurrency.lockutils [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.646300] env[62692]: DEBUG oslo_concurrency.lockutils [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 928.646679] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-acd9541e-7045-4124-ab4b-2c56dabdeae2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.652766] env[62692]: DEBUG oslo_vmware.api [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Waiting for the task: (returnval){ [ 928.652766] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52eed7f5-f2be-6124-9a15-d4773641c326" [ 928.652766] env[62692]: _type = "Task" [ 928.652766] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.662476] env[62692]: DEBUG oslo_vmware.api [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52eed7f5-f2be-6124-9a15-d4773641c326, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.710800] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]521f4366-77a6-ee61-89f2-7d1d41de4df8, 'name': SearchDatastore_Task, 'duration_secs': 0.028864} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.712162] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.712162] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] fac426b6-747a-455e-85a1-44e71b432180/fac426b6-747a-455e-85a1-44e71b432180.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 928.712162] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2ba5e0ca-41bb-47f3-8907-de267a0acfc6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.724905] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141377, 'name': Rename_Task, 'duration_secs': 0.32869} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.726436] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 928.726677] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 928.726677] env[62692]: value = "task-1141381" [ 928.726677] env[62692]: _type = "Task" [ 928.726677] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.726875] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f773965a-6bd7-4b6e-86b5-a8fc311dfa7a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.740547] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141381, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.740547] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 928.740547] env[62692]: value = "task-1141382" [ 928.740547] env[62692]: _type = "Task" [ 928.740547] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.755020] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141382, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.900928] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141378, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065529} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.902038] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 928.903053] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c558304-79e7-4981-b1de-51e17f242055 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.930711] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] b23d3484-4fc5-46b2-8e8c-bdaab087f046/b23d3484-4fc5-46b2-8e8c-bdaab087f046.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 928.931863] env[62692]: DEBUG nova.scheduler.client.report [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Updated inventory for provider 470758a8-5efa-4875-aad5-f512a727752f with generation 98 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 928.932120] env[62692]: DEBUG nova.compute.provider_tree [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Updating resource provider 470758a8-5efa-4875-aad5-f512a727752f generation from 98 to 99 during operation: update_inventory {{(pid=62692) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 928.932311] env[62692]: DEBUG nova.compute.provider_tree [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 928.935667] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dadf211b-5473-46ef-a787-16de874be585 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.951201] env[62692]: DEBUG oslo_concurrency.lockutils [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.657s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.953145] env[62692]: DEBUG oslo_concurrency.lockutils [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 25.435s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.961960] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 928.961960] env[62692]: value = "task-1141383" [ 928.961960] env[62692]: _type = "Task" [ 928.961960] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.973643] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141383, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.989545] env[62692]: INFO nova.scheduler.client.report [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Deleted allocations for instance 433f740c-bde4-4d33-baca-5a7ac8179da1 [ 928.999658] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141379, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.112150] env[62692]: DEBUG oslo_vmware.api [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141380, 'name': PowerOffVM_Task, 'duration_secs': 0.23333} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.112538] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 929.112752] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Volume detach. Driver type: vmdk {{(pid=62692) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 929.112976] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-248983', 'volume_id': '14bc011a-54f2-4452-95c6-b9a4f2020742', 'name': 'volume-14bc011a-54f2-4452-95c6-b9a4f2020742', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3fd2959b-80cb-470c-8c2a-40b7630458ed', 'attached_at': '', 'detached_at': '', 'volume_id': '14bc011a-54f2-4452-95c6-b9a4f2020742', 'serial': '14bc011a-54f2-4452-95c6-b9a4f2020742'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 929.113865] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b40d58d-abdb-420b-a505-813fe0d194a9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.148612] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b9be75b-209a-404d-8ffe-99f56bed0ce1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.161194] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6debde2-ef8d-4fd7-b846-4e12543f8006 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.170169] env[62692]: DEBUG oslo_vmware.api [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52eed7f5-f2be-6124-9a15-d4773641c326, 'name': SearchDatastore_Task, 'duration_secs': 0.011084} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.186073] env[62692]: DEBUG oslo_concurrency.lockutils [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.186378] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 929.186621] env[62692]: DEBUG oslo_concurrency.lockutils [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.186770] env[62692]: DEBUG oslo_concurrency.lockutils [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.186961] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 929.188318] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fd1dec89-d8cf-4cd0-9346-5511dad6b29b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.191029] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d48870c-1657-4326-a443-afc8fa5917b0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.213929] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] The volume has not been displaced from its original location: [datastore1] volume-14bc011a-54f2-4452-95c6-b9a4f2020742/volume-14bc011a-54f2-4452-95c6-b9a4f2020742.vmdk. No consolidation needed. {{(pid=62692) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 929.219449] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Reconfiguring VM instance instance-00000043 to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 929.221036] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c05e05c-c45a-45e1-aabc-f8c2bcb68bce {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.235435] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 929.235682] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 929.236556] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14ff91fb-4d7a-41bc-886e-a5cc64691dec {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.248522] env[62692]: DEBUG oslo_vmware.api [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Waiting for the task: (returnval){ [ 929.248522] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52025c6d-0083-3c39-0f31-95a08048f4ca" [ 929.248522] env[62692]: _type = "Task" [ 929.248522] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.249824] env[62692]: DEBUG oslo_vmware.api [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 929.249824] env[62692]: value = "task-1141384" [ 929.249824] env[62692]: _type = "Task" [ 929.249824] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.263201] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141382, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.263201] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141381, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.271597] env[62692]: DEBUG oslo_vmware.api [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52025c6d-0083-3c39-0f31-95a08048f4ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.275640] env[62692]: DEBUG oslo_vmware.api [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141384, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.457028] env[62692]: DEBUG nova.objects.instance [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Lazy-loading 'migration_context' on Instance uuid 2ccb76cf-f641-4306-a137-fb2417285df9 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 929.475344] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141383, 'name': ReconfigVM_Task, 'duration_secs': 0.432258} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.476488] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Reconfigured VM instance instance-00000048 to attach disk [datastore2] b23d3484-4fc5-46b2-8e8c-bdaab087f046/b23d3484-4fc5-46b2-8e8c-bdaab087f046.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 929.477251] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-26e4874d-de9c-446f-9bed-4332e56e39e5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.489300] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 929.489300] env[62692]: value = "task-1141385" [ 929.489300] env[62692]: _type = "Task" [ 929.489300] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.498130] env[62692]: DEBUG oslo_vmware.api [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141379, 'name': PowerOnVM_Task, 'duration_secs': 0.711394} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.498480] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 929.499025] env[62692]: INFO nova.compute.manager [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Took 18.43 seconds to spawn the instance on the hypervisor. [ 929.499025] env[62692]: DEBUG nova.compute.manager [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 929.499706] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a249a1c-f7e7-4339-a660-29f38b373a8a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.508491] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141385, 'name': Rename_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.509574] env[62692]: DEBUG oslo_concurrency.lockutils [None req-771f001c-0bb9-4e1e-a5f4-b16efeaa2469 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "433f740c-bde4-4d33-baca-5a7ac8179da1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.725s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.750356] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141381, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.590026} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.750356] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] fac426b6-747a-455e-85a1-44e71b432180/fac426b6-747a-455e-85a1-44e71b432180.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 929.750356] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 929.753496] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-876f1931-ee14-4cdb-8ac8-1b10a77b4192 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.757912] env[62692]: DEBUG oslo_vmware.api [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141382, 'name': PowerOnVM_Task, 'duration_secs': 0.76309} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.758906] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 929.759060] env[62692]: INFO nova.compute.manager [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Took 10.68 seconds to spawn the instance on the hypervisor. [ 929.759297] env[62692]: DEBUG nova.compute.manager [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 929.760102] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50056a93-0bf0-4a19-8762-96dfdb94eb9e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.766599] env[62692]: DEBUG oslo_vmware.api [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52025c6d-0083-3c39-0f31-95a08048f4ca, 'name': SearchDatastore_Task, 'duration_secs': 0.047224} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.772056] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 929.772056] env[62692]: value = "task-1141386" [ 929.772056] env[62692]: _type = "Task" [ 929.772056] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.772371] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b43096aa-5422-451b-9902-f769573c95a2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.783315] env[62692]: DEBUG oslo_vmware.api [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141384, 'name': ReconfigVM_Task, 'duration_secs': 0.235279} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.788749] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Reconfigured VM instance instance-00000043 to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 929.794029] env[62692]: DEBUG oslo_vmware.api [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Waiting for the task: (returnval){ [ 929.794029] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]527fb925-3691-4798-93aa-ff630623305c" [ 929.794029] env[62692]: _type = "Task" [ 929.794029] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.794550] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c8e0f29-7d10-4551-9fff-0dcf93cc7bd5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.811381] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141386, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.821083] env[62692]: DEBUG oslo_vmware.api [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]527fb925-3691-4798-93aa-ff630623305c, 'name': SearchDatastore_Task, 'duration_secs': 0.016409} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.822795] env[62692]: DEBUG oslo_concurrency.lockutils [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.823214] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 8a56e6df-eea1-41f4-9360-4f06d2f516a3/8a56e6df-eea1-41f4-9360-4f06d2f516a3.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 929.823667] env[62692]: DEBUG oslo_vmware.api [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 929.823667] env[62692]: value = "task-1141387" [ 929.823667] env[62692]: _type = "Task" [ 929.823667] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.824757] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f01ede9a-a909-4063-994e-4efd125ce2a8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.838938] env[62692]: DEBUG oslo_vmware.api [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141387, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.840289] env[62692]: DEBUG oslo_vmware.api [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Waiting for the task: (returnval){ [ 929.840289] env[62692]: value = "task-1141388" [ 929.840289] env[62692]: _type = "Task" [ 929.840289] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.849067] env[62692]: DEBUG oslo_vmware.api [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Task: {'id': task-1141388, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.001701] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141385, 'name': Rename_Task, 'duration_secs': 0.171207} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.001887] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 930.002132] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-62b10324-ce6d-48fb-9d3c-4cc7aa2b50a0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.009970] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 930.009970] env[62692]: value = "task-1141389" [ 930.009970] env[62692]: _type = "Task" [ 930.009970] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.022228] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141389, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.027149] env[62692]: INFO nova.compute.manager [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Took 55.60 seconds to build instance. [ 930.287867] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141386, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072592} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.288176] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 930.288950] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04c9650f-eb91-4ae3-b7fa-8d8a13351ee6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.317730] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] fac426b6-747a-455e-85a1-44e71b432180/fac426b6-747a-455e-85a1-44e71b432180.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 930.322283] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-281900e5-f23c-40d3-b59a-f99be088c0d2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.337186] env[62692]: INFO nova.compute.manager [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Took 46.83 seconds to build instance. [ 930.348028] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 930.348028] env[62692]: value = "task-1141390" [ 930.348028] env[62692]: _type = "Task" [ 930.348028] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.351901] env[62692]: DEBUG oslo_vmware.api [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141387, 'name': ReconfigVM_Task, 'duration_secs': 0.493923} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.358352] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-248983', 'volume_id': '14bc011a-54f2-4452-95c6-b9a4f2020742', 'name': 'volume-14bc011a-54f2-4452-95c6-b9a4f2020742', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3fd2959b-80cb-470c-8c2a-40b7630458ed', 'attached_at': '', 'detached_at': '', 'volume_id': '14bc011a-54f2-4452-95c6-b9a4f2020742', 'serial': '14bc011a-54f2-4452-95c6-b9a4f2020742'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 930.358895] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 930.363371] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba7eeb5b-5e93-4c53-94f9-7d8c49c41f73 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.366168] env[62692]: DEBUG oslo_vmware.api [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Task: {'id': task-1141388, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.378289] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 930.378600] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141390, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.378978] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-52bb61dc-7906-4a17-8ef8-8a5cb2e47981 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.423072] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c8b0c96-d176-478a-86ff-95a80bf92b68 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.432651] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83f57a06-03a2-4237-a437-035d83951d20 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.484935] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73bd3bcd-e18c-478e-9095-54cebd05459e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.489631] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 930.490170] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 930.490577] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Deleting the datastore file [datastore1] 3fd2959b-80cb-470c-8c2a-40b7630458ed {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 930.491026] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a2be1c94-573d-4256-91c6-11ac7ce686c7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.506088] env[62692]: DEBUG oslo_vmware.api [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 930.506088] env[62692]: value = "task-1141392" [ 930.506088] env[62692]: _type = "Task" [ 930.506088] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.508317] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-330011dc-6b7c-49ce-a3b0-a587dbc22686 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.528059] env[62692]: DEBUG oslo_vmware.api [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141392, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.541655] env[62692]: DEBUG oslo_concurrency.lockutils [None req-baa63bf8-10ac-4d34-96f8-26d191b25c8f tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "44699df6-5262-4700-a096-8c0e450323d2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.124s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.542282] env[62692]: DEBUG nova.compute.provider_tree [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 930.543750] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141389, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.842325] env[62692]: DEBUG oslo_concurrency.lockutils [None req-13b6d62b-f292-4f88-a117-3b00261fc424 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.870s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.851695] env[62692]: DEBUG oslo_vmware.api [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Task: {'id': task-1141388, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.943504} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.852117] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 8a56e6df-eea1-41f4-9360-4f06d2f516a3/8a56e6df-eea1-41f4-9360-4f06d2f516a3.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 930.852388] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 930.855772] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b292d592-39d6-4067-acf8-fb5f4c17cd4d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.862832] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141390, 'name': ReconfigVM_Task, 'duration_secs': 0.379751} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.864143] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Reconfigured VM instance instance-00000046 to attach disk [datastore2] fac426b6-747a-455e-85a1-44e71b432180/fac426b6-747a-455e-85a1-44e71b432180.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 930.864947] env[62692]: DEBUG oslo_vmware.api [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Waiting for the task: (returnval){ [ 930.864947] env[62692]: value = "task-1141393" [ 930.864947] env[62692]: _type = "Task" [ 930.864947] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.865162] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-826639e1-3915-4ca6-9a3a-776ce6a6fbdd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.876487] env[62692]: DEBUG oslo_vmware.api [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Task: {'id': task-1141393, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.877899] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 930.877899] env[62692]: value = "task-1141394" [ 930.877899] env[62692]: _type = "Task" [ 930.877899] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.891070] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141394, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.018419] env[62692]: DEBUG oslo_vmware.api [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141392, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.443101} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.021633] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 931.021817] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 931.021993] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 931.022205] env[62692]: INFO nova.compute.manager [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Took 2.43 seconds to destroy the instance on the hypervisor. [ 931.022466] env[62692]: DEBUG oslo.service.loopingcall [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 931.022667] env[62692]: DEBUG nova.compute.manager [-] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 931.022760] env[62692]: DEBUG nova.network.neutron [-] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 931.030225] env[62692]: DEBUG oslo_vmware.api [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141389, 'name': PowerOnVM_Task, 'duration_secs': 0.749669} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.030474] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 931.030667] env[62692]: INFO nova.compute.manager [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Took 8.83 seconds to spawn the instance on the hypervisor. [ 931.030842] env[62692]: DEBUG nova.compute.manager [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 931.031652] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb62a215-2804-4bda-94b3-69706a839aac {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.045885] env[62692]: DEBUG nova.scheduler.client.report [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 931.049038] env[62692]: DEBUG nova.compute.manager [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 931.144380] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "44699df6-5262-4700-a096-8c0e450323d2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.144581] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "44699df6-5262-4700-a096-8c0e450323d2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.144799] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "44699df6-5262-4700-a096-8c0e450323d2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.144986] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "44699df6-5262-4700-a096-8c0e450323d2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.145178] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "44699df6-5262-4700-a096-8c0e450323d2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.147365] env[62692]: INFO nova.compute.manager [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Terminating instance [ 931.150670] env[62692]: DEBUG nova.compute.manager [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 931.150885] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 931.153487] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e379bde2-01f1-46cc-a24e-0bcc4dfe6caa {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.162159] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 931.162410] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1c7dcc9c-3903-4284-9735-6bd471dbbeb4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.170155] env[62692]: DEBUG oslo_vmware.api [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 931.170155] env[62692]: value = "task-1141395" [ 931.170155] env[62692]: _type = "Task" [ 931.170155] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.179044] env[62692]: DEBUG oslo_vmware.api [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141395, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.215303] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6dfa270a-83a7-4eda-a4cd-5f974faf09ce tempest-ServersAdminTestJSON-1405005272 tempest-ServersAdminTestJSON-1405005272-project-admin] Acquiring lock "refresh_cache-0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.215561] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6dfa270a-83a7-4eda-a4cd-5f974faf09ce tempest-ServersAdminTestJSON-1405005272 tempest-ServersAdminTestJSON-1405005272-project-admin] Acquired lock "refresh_cache-0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.215665] env[62692]: DEBUG nova.network.neutron [None req-6dfa270a-83a7-4eda-a4cd-5f974faf09ce tempest-ServersAdminTestJSON-1405005272 tempest-ServersAdminTestJSON-1405005272-project-admin] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 931.386293] env[62692]: DEBUG oslo_vmware.api [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Task: {'id': task-1141393, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.156853} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.391695] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 931.393719] env[62692]: DEBUG nova.compute.manager [req-df41aca1-5d2b-4869-af99-0f670d56bcd0 req-a9c58709-7bea-43f2-8aec-3f229b80f66e service nova] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Received event network-vif-deleted-2d5b5313-b87f-4cb1-9478-6fbbae387430 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 931.393958] env[62692]: INFO nova.compute.manager [req-df41aca1-5d2b-4869-af99-0f670d56bcd0 req-a9c58709-7bea-43f2-8aec-3f229b80f66e service nova] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Neutron deleted interface 2d5b5313-b87f-4cb1-9478-6fbbae387430; detaching it from the instance and deleting it from the info cache [ 931.394106] env[62692]: DEBUG nova.network.neutron [req-df41aca1-5d2b-4869-af99-0f670d56bcd0 req-a9c58709-7bea-43f2-8aec-3f229b80f66e service nova] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.395933] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80ce5474-c19d-44a7-a29c-b748b022d288 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.405654] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141394, 'name': Rename_Task, 'duration_secs': 0.188209} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.422420] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 931.432969] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] 8a56e6df-eea1-41f4-9360-4f06d2f516a3/8a56e6df-eea1-41f4-9360-4f06d2f516a3.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 931.433336] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b6800338-b805-4433-8a14-c2ec3b14a6ec {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.435763] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e4918bbd-2b01-4ab2-9715-28ad43c28a48 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.457556] env[62692]: DEBUG oslo_vmware.api [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Waiting for the task: (returnval){ [ 931.457556] env[62692]: value = "task-1141397" [ 931.457556] env[62692]: _type = "Task" [ 931.457556] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.457847] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 931.457847] env[62692]: value = "task-1141396" [ 931.457847] env[62692]: _type = "Task" [ 931.457847] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.476334] env[62692]: DEBUG oslo_vmware.api [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Task: {'id': task-1141397, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.483214] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141396, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.555068] env[62692]: INFO nova.compute.manager [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Took 43.12 seconds to build instance. [ 931.588806] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.688876] env[62692]: DEBUG oslo_vmware.api [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141395, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.853087] env[62692]: DEBUG nova.network.neutron [-] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.900025] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-34d54907-42ec-4ac4-9665-3c14f324776e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.918024] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f9ad8b8-8e7b-4f5e-b800-b58b2049daac {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.957183] env[62692]: DEBUG nova.compute.manager [req-df41aca1-5d2b-4869-af99-0f670d56bcd0 req-a9c58709-7bea-43f2-8aec-3f229b80f66e service nova] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Detach interface failed, port_id=2d5b5313-b87f-4cb1-9478-6fbbae387430, reason: Instance 3fd2959b-80cb-470c-8c2a-40b7630458ed could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 931.973677] env[62692]: DEBUG oslo_vmware.api [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141396, 'name': PowerOnVM_Task, 'duration_secs': 0.479114} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.977748] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 931.978117] env[62692]: DEBUG nova.compute.manager [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 931.978485] env[62692]: DEBUG oslo_vmware.api [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Task: {'id': task-1141397, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.979412] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aeca7e6-0b4f-438b-ba9b-85b2bd1c0655 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.071026] env[62692]: DEBUG oslo_concurrency.lockutils [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 3.114s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.074318] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e460f4fe-c82c-49cf-a78a-65af5233bebb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "b23d3484-4fc5-46b2-8e8c-bdaab087f046" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.701s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.080033] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.713s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.080033] env[62692]: INFO nova.compute.claims [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 932.187812] env[62692]: DEBUG oslo_vmware.api [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141395, 'name': PowerOffVM_Task, 'duration_secs': 0.573983} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.192020] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 932.192020] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 932.192020] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5d833467-37bc-4d37-b973-d5109c4f8787 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.257048] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 932.257048] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 932.257048] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Deleting the datastore file [datastore1] 44699df6-5262-4700-a096-8c0e450323d2 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 932.257454] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b4a4b54d-81fa-4671-8b37-0acc540b8a39 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.261389] env[62692]: DEBUG nova.network.neutron [None req-6dfa270a-83a7-4eda-a4cd-5f974faf09ce tempest-ServersAdminTestJSON-1405005272 tempest-ServersAdminTestJSON-1405005272-project-admin] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Updating instance_info_cache with network_info: [{"id": "0047b080-5217-4718-a134-a87f4556102b", "address": "fa:16:3e:0d:46:60", "network": {"id": "0e06e04a-5213-4d06-b3f1-b2b8f437e6d0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1533469056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "266c39fe585d41079d0278ecd1d1c6a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "678ebbe4-4c53-4eaf-a689-93981310f37d", "external-id": "nsx-vlan-transportzone-443", "segmentation_id": 443, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0047b080-52", "ovs_interfaceid": "0047b080-5217-4718-a134-a87f4556102b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.268025] env[62692]: DEBUG oslo_vmware.api [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 932.268025] env[62692]: value = "task-1141399" [ 932.268025] env[62692]: _type = "Task" [ 932.268025] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.276087] env[62692]: DEBUG oslo_vmware.api [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141399, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.356693] env[62692]: INFO nova.compute.manager [-] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Took 1.33 seconds to deallocate network for instance. [ 932.470758] env[62692]: DEBUG oslo_vmware.api [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Task: {'id': task-1141397, 'name': ReconfigVM_Task, 'duration_secs': 0.640657} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.471142] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Reconfigured VM instance instance-00000049 to attach disk [datastore2] 8a56e6df-eea1-41f4-9360-4f06d2f516a3/8a56e6df-eea1-41f4-9360-4f06d2f516a3.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 932.472798] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6a3e6b11-fd54-4c86-abca-0d945c7a4615 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.483362] env[62692]: DEBUG oslo_vmware.api [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Waiting for the task: (returnval){ [ 932.483362] env[62692]: value = "task-1141400" [ 932.483362] env[62692]: _type = "Task" [ 932.483362] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.507014] env[62692]: DEBUG oslo_vmware.api [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Task: {'id': task-1141400, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.514489] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.765028] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6dfa270a-83a7-4eda-a4cd-5f974faf09ce tempest-ServersAdminTestJSON-1405005272 tempest-ServersAdminTestJSON-1405005272-project-admin] Releasing lock "refresh_cache-0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.765315] env[62692]: DEBUG nova.compute.manager [None req-6dfa270a-83a7-4eda-a4cd-5f974faf09ce tempest-ServersAdminTestJSON-1405005272 tempest-ServersAdminTestJSON-1405005272-project-admin] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Inject network info {{(pid=62692) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 932.767715] env[62692]: DEBUG nova.compute.manager [None req-6dfa270a-83a7-4eda-a4cd-5f974faf09ce tempest-ServersAdminTestJSON-1405005272 tempest-ServersAdminTestJSON-1405005272-project-admin] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] network_info to inject: |[{"id": "0047b080-5217-4718-a134-a87f4556102b", "address": "fa:16:3e:0d:46:60", "network": {"id": "0e06e04a-5213-4d06-b3f1-b2b8f437e6d0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1533469056-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "266c39fe585d41079d0278ecd1d1c6a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "678ebbe4-4c53-4eaf-a689-93981310f37d", "external-id": "nsx-vlan-transportzone-443", "segmentation_id": 443, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0047b080-52", "ovs_interfaceid": "0047b080-5217-4718-a134-a87f4556102b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 932.774227] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6dfa270a-83a7-4eda-a4cd-5f974faf09ce tempest-ServersAdminTestJSON-1405005272 tempest-ServersAdminTestJSON-1405005272-project-admin] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Reconfiguring VM instance to set the machine id {{(pid=62692) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1802}} [ 932.774643] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53fb2af1-5cf3-4b48-8576-5640d9420f04 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.799970] env[62692]: DEBUG oslo_vmware.api [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141399, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.398505} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.802117] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 932.802117] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 932.802117] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 932.802117] env[62692]: INFO nova.compute.manager [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Took 1.65 seconds to destroy the instance on the hypervisor. [ 932.802319] env[62692]: DEBUG oslo.service.loopingcall [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 932.802858] env[62692]: DEBUG oslo_vmware.api [None req-6dfa270a-83a7-4eda-a4cd-5f974faf09ce tempest-ServersAdminTestJSON-1405005272 tempest-ServersAdminTestJSON-1405005272-project-admin] Waiting for the task: (returnval){ [ 932.802858] env[62692]: value = "task-1141401" [ 932.802858] env[62692]: _type = "Task" [ 932.802858] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.802858] env[62692]: DEBUG nova.compute.manager [-] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 932.802858] env[62692]: DEBUG nova.network.neutron [-] [instance: 44699df6-5262-4700-a096-8c0e450323d2] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 932.816836] env[62692]: DEBUG oslo_vmware.api [None req-6dfa270a-83a7-4eda-a4cd-5f974faf09ce tempest-ServersAdminTestJSON-1405005272 tempest-ServersAdminTestJSON-1405005272-project-admin] Task: {'id': task-1141401, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.925560] env[62692]: INFO nova.compute.manager [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Took 0.57 seconds to detach 1 volumes for instance. [ 932.992560] env[62692]: DEBUG oslo_vmware.api [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Task: {'id': task-1141400, 'name': Rename_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.317306] env[62692]: DEBUG oslo_vmware.api [None req-6dfa270a-83a7-4eda-a4cd-5f974faf09ce tempest-ServersAdminTestJSON-1405005272 tempest-ServersAdminTestJSON-1405005272-project-admin] Task: {'id': task-1141401, 'name': ReconfigVM_Task, 'duration_secs': 0.39533} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.320647] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6dfa270a-83a7-4eda-a4cd-5f974faf09ce tempest-ServersAdminTestJSON-1405005272 tempest-ServersAdminTestJSON-1405005272-project-admin] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Reconfigured VM instance to set the machine id {{(pid=62692) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1805}} [ 933.327822] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquiring lock "fac426b6-747a-455e-85a1-44e71b432180" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.328177] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Lock "fac426b6-747a-455e-85a1-44e71b432180" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.328443] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquiring lock "fac426b6-747a-455e-85a1-44e71b432180-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.328692] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Lock "fac426b6-747a-455e-85a1-44e71b432180-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.328884] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Lock "fac426b6-747a-455e-85a1-44e71b432180-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.333433] env[62692]: INFO nova.compute.manager [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Terminating instance [ 933.336701] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquiring lock "refresh_cache-fac426b6-747a-455e-85a1-44e71b432180" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.336701] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquired lock "refresh_cache-fac426b6-747a-455e-85a1-44e71b432180" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.336701] env[62692]: DEBUG nova.network.neutron [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 933.434933] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.500614] env[62692]: DEBUG oslo_vmware.api [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Task: {'id': task-1141400, 'name': Rename_Task, 'duration_secs': 0.768485} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.501442] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 933.501543] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9117b8b4-df61-41d6-900a-55787b5b8d17 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.517590] env[62692]: DEBUG oslo_vmware.api [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Waiting for the task: (returnval){ [ 933.517590] env[62692]: value = "task-1141402" [ 933.517590] env[62692]: _type = "Task" [ 933.517590] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.536062] env[62692]: DEBUG oslo_vmware.api [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Task: {'id': task-1141402, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.623521] env[62692]: INFO nova.compute.manager [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Swapping old allocation on dict_keys(['470758a8-5efa-4875-aad5-f512a727752f']) held by migration 82adcbf0-fa69-4f63-b376-ff031ef9c49e for instance [ 933.643851] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6da5280c-c41b-48b3-9863-2efe9cbd57c9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.653150] env[62692]: DEBUG nova.scheduler.client.report [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Overwriting current allocation {'allocations': {'470758a8-5efa-4875-aad5-f512a727752f': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 100}}, 'project_id': '43057671f72b4bfa83f33855a1bf4422', 'user_id': 'd33baeebac3d4fad907677f763bf71be', 'consumer_generation': 1} on consumer 2ccb76cf-f641-4306-a137-fb2417285df9 {{(pid=62692) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 933.657894] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d97bc4d3-6ce1-4d6e-b289-39a88b6fd7f7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.695131] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4af40629-8406-49f8-bb0d-97c73398c800 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.706384] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff7251f5-1f0c-443d-9a75-efe3b4af34fe {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.728707] env[62692]: DEBUG nova.compute.provider_tree [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 933.781990] env[62692]: DEBUG nova.compute.manager [req-6ac5096e-fae9-4c11-9ad5-395b3b6e7dc4 req-fe7f7c65-ef23-4098-b97b-486b7dc9cb71 service nova] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Received event network-vif-deleted-6702eb8d-7d7d-4384-8eb4-d2196df920ef {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 933.781990] env[62692]: INFO nova.compute.manager [req-6ac5096e-fae9-4c11-9ad5-395b3b6e7dc4 req-fe7f7c65-ef23-4098-b97b-486b7dc9cb71 service nova] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Neutron deleted interface 6702eb8d-7d7d-4384-8eb4-d2196df920ef; detaching it from the instance and deleting it from the info cache [ 933.782320] env[62692]: DEBUG nova.network.neutron [req-6ac5096e-fae9-4c11-9ad5-395b3b6e7dc4 req-fe7f7c65-ef23-4098-b97b-486b7dc9cb71 service nova] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.809924] env[62692]: DEBUG oslo_concurrency.lockutils [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquiring lock "refresh_cache-2ccb76cf-f641-4306-a137-fb2417285df9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.810174] env[62692]: DEBUG oslo_concurrency.lockutils [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquired lock "refresh_cache-2ccb76cf-f641-4306-a137-fb2417285df9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.810351] env[62692]: DEBUG nova.network.neutron [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 933.869673] env[62692]: DEBUG nova.network.neutron [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 933.997565] env[62692]: DEBUG nova.network.neutron [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.002563] env[62692]: DEBUG oslo_concurrency.lockutils [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "f7664593-1b7e-495f-8d0e-9137d7f180cc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.002801] env[62692]: DEBUG oslo_concurrency.lockutils [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "f7664593-1b7e-495f-8d0e-9137d7f180cc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.028640] env[62692]: DEBUG oslo_vmware.api [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Task: {'id': task-1141402, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.117988] env[62692]: DEBUG nova.network.neutron [-] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.232680] env[62692]: DEBUG nova.scheduler.client.report [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 934.289018] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7567b631-86db-451d-80fd-9460e9693111 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.297901] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00058d0b-6d69-4583-89d2-2f3173114974 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.340038] env[62692]: DEBUG nova.compute.manager [req-6ac5096e-fae9-4c11-9ad5-395b3b6e7dc4 req-fe7f7c65-ef23-4098-b97b-486b7dc9cb71 service nova] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Detach interface failed, port_id=6702eb8d-7d7d-4384-8eb4-d2196df920ef, reason: Instance 44699df6-5262-4700-a096-8c0e450323d2 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 934.506797] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Releasing lock "refresh_cache-fac426b6-747a-455e-85a1-44e71b432180" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.506797] env[62692]: DEBUG nova.compute.manager [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 934.506797] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 934.506797] env[62692]: DEBUG nova.compute.manager [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 934.510264] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bec111dc-0e96-42ef-8cfb-8620080622fb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.525299] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 934.530012] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-26fb5594-c7b6-49d3-906c-606bd1fa7556 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.543260] env[62692]: DEBUG oslo_vmware.api [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Task: {'id': task-1141402, 'name': PowerOnVM_Task, 'duration_secs': 0.750676} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.543370] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 934.543517] env[62692]: INFO nova.compute.manager [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Took 9.79 seconds to spawn the instance on the hypervisor. [ 934.543697] env[62692]: DEBUG nova.compute.manager [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 934.544067] env[62692]: DEBUG oslo_vmware.api [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 934.544067] env[62692]: value = "task-1141403" [ 934.544067] env[62692]: _type = "Task" [ 934.544067] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.544825] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5efa4a3d-b3da-4173-9c1d-92e1e3fe4264 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.564951] env[62692]: DEBUG oslo_vmware.api [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141403, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.596658] env[62692]: DEBUG nova.network.neutron [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Updating instance_info_cache with network_info: [{"id": "951b0add-cfaa-4506-a6ac-56bcb8d1041e", "address": "fa:16:3e:70:90:85", "network": {"id": "a74c4d30-3568-4e50-98f5-d2fcc05ec728", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.89", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "989968e8796a4a70a4dbd045418a475f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap951b0add-cf", "ovs_interfaceid": "951b0add-cfaa-4506-a6ac-56bcb8d1041e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.619375] env[62692]: INFO nova.compute.manager [-] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Took 1.82 seconds to deallocate network for instance. [ 934.740839] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.664s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.741472] env[62692]: DEBUG nova.compute.manager [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 934.744598] env[62692]: DEBUG oslo_concurrency.lockutils [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.327s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.746073] env[62692]: INFO nova.compute.claims [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 935.034690] env[62692]: DEBUG oslo_concurrency.lockutils [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.057864] env[62692]: DEBUG oslo_vmware.api [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141403, 'name': PowerOffVM_Task, 'duration_secs': 0.142344} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.058165] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 935.058345] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 935.058586] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-84ffdd9e-5976-4579-a237-8c168ad5c13a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.080041] env[62692]: INFO nova.compute.manager [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Took 40.53 seconds to build instance. [ 935.088102] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 935.088793] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 935.088847] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Deleting the datastore file [datastore2] fac426b6-747a-455e-85a1-44e71b432180 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 935.089249] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-82cf7ba0-7186-4b82-a3d6-a8fc420b865c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.102191] env[62692]: DEBUG oslo_concurrency.lockutils [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Releasing lock "refresh_cache-2ccb76cf-f641-4306-a137-fb2417285df9" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.102191] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 935.102862] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4332e9f3-a27a-44e5-a83b-c94448f80297 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.104590] env[62692]: DEBUG oslo_vmware.api [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 935.104590] env[62692]: value = "task-1141405" [ 935.104590] env[62692]: _type = "Task" [ 935.104590] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.110909] env[62692]: DEBUG oslo_vmware.api [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Waiting for the task: (returnval){ [ 935.110909] env[62692]: value = "task-1141406" [ 935.110909] env[62692]: _type = "Task" [ 935.110909] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.116784] env[62692]: DEBUG oslo_vmware.api [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141405, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.124381] env[62692]: DEBUG oslo_vmware.api [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141406, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.129640] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.252129] env[62692]: DEBUG nova.compute.utils [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 935.252792] env[62692]: DEBUG nova.compute.manager [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 935.256221] env[62692]: DEBUG nova.network.neutron [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 935.297327] env[62692]: INFO nova.compute.manager [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Rebuilding instance [ 935.302395] env[62692]: DEBUG nova.policy [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fcab56b7513648d7a991cb5c426042bf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '54c19ee28b1c4b829ffa32a61a4d22a9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 935.363374] env[62692]: DEBUG nova.compute.manager [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 935.364343] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a6b2a92-e096-49d0-8cb7-70647659c8f1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.582826] env[62692]: DEBUG oslo_concurrency.lockutils [None req-778d12c8-cb85-43f7-82fc-566e01cf5fd2 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Lock "8a56e6df-eea1-41f4-9360-4f06d2f516a3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.309s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.617209] env[62692]: DEBUG oslo_vmware.api [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141405, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.175908} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.623635] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 935.623856] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 935.624051] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 935.624237] env[62692]: INFO nova.compute.manager [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Took 1.12 seconds to destroy the instance on the hypervisor. [ 935.624489] env[62692]: DEBUG oslo.service.loopingcall [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 935.624700] env[62692]: DEBUG nova.compute.manager [-] [instance: fac426b6-747a-455e-85a1-44e71b432180] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 935.624794] env[62692]: DEBUG nova.network.neutron [-] [instance: fac426b6-747a-455e-85a1-44e71b432180] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 935.635555] env[62692]: DEBUG oslo_vmware.api [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141406, 'name': PowerOffVM_Task, 'duration_secs': 0.295942} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.635869] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 935.636708] env[62692]: DEBUG nova.virt.hardware [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:32:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='a53d9694-cbf0-4c37-a7c0-aa52b2363bc3',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1692421942',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 935.636962] env[62692]: DEBUG nova.virt.hardware [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 935.637938] env[62692]: DEBUG nova.virt.hardware [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 935.638662] env[62692]: DEBUG nova.virt.hardware [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 935.639385] env[62692]: DEBUG nova.virt.hardware [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 935.639793] env[62692]: DEBUG nova.virt.hardware [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 935.640338] env[62692]: DEBUG nova.virt.hardware [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 935.640568] env[62692]: DEBUG nova.virt.hardware [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 935.640779] env[62692]: DEBUG nova.virt.hardware [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 935.640963] env[62692]: DEBUG nova.virt.hardware [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 935.641184] env[62692]: DEBUG nova.virt.hardware [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 935.647731] env[62692]: DEBUG nova.network.neutron [-] [instance: fac426b6-747a-455e-85a1-44e71b432180] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 935.649123] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c2d614d-fe7d-4daf-8c5e-22256b45481b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.663949] env[62692]: DEBUG nova.network.neutron [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Successfully created port: 62803691-31f7-422f-b373-a85721fe22b6 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 935.672830] env[62692]: DEBUG oslo_vmware.api [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Waiting for the task: (returnval){ [ 935.672830] env[62692]: value = "task-1141407" [ 935.672830] env[62692]: _type = "Task" [ 935.672830] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.684452] env[62692]: DEBUG oslo_vmware.api [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141407, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.760041] env[62692]: DEBUG nova.compute.manager [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 935.879596] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 935.879908] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6280c37d-941c-4105-96f4-20a2b061a915 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.889394] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 935.889394] env[62692]: value = "task-1141408" [ 935.889394] env[62692]: _type = "Task" [ 935.889394] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.907318] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141408, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.947163] env[62692]: DEBUG oslo_concurrency.lockutils [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Acquiring lock "16790505-cdf6-4937-8839-fa685a5f413e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.947163] env[62692]: DEBUG oslo_concurrency.lockutils [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Lock "16790505-cdf6-4937-8839-fa685a5f413e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.163625] env[62692]: DEBUG nova.network.neutron [-] [instance: fac426b6-747a-455e-85a1-44e71b432180] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.189039] env[62692]: DEBUG oslo_vmware.api [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141407, 'name': ReconfigVM_Task, 'duration_secs': 0.27171} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.193868] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6712c84-2b35-400d-80a9-f88139e56b15 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.215674] env[62692]: DEBUG nova.virt.hardware [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:32:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='a53d9694-cbf0-4c37-a7c0-aa52b2363bc3',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1692421942',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 936.216187] env[62692]: DEBUG nova.virt.hardware [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 936.216187] env[62692]: DEBUG nova.virt.hardware [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 936.216327] env[62692]: DEBUG nova.virt.hardware [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 936.216430] env[62692]: DEBUG nova.virt.hardware [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 936.216580] env[62692]: DEBUG nova.virt.hardware [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 936.216790] env[62692]: DEBUG nova.virt.hardware [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 936.216956] env[62692]: DEBUG nova.virt.hardware [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 936.217145] env[62692]: DEBUG nova.virt.hardware [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 936.217314] env[62692]: DEBUG nova.virt.hardware [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 936.217621] env[62692]: DEBUG nova.virt.hardware [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 936.221542] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83a6273f-1340-4276-91f6-6f03acf8e2f5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.229234] env[62692]: DEBUG oslo_vmware.api [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Waiting for the task: (returnval){ [ 936.229234] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]522c02e0-2190-a4ab-fc51-270d1b500295" [ 936.229234] env[62692]: _type = "Task" [ 936.229234] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.243220] env[62692]: DEBUG oslo_vmware.api [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]522c02e0-2190-a4ab-fc51-270d1b500295, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.289394] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5224cc9-d89f-4353-a5d6-555207f0af84 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.297695] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be1928c9-c576-4996-aaa8-20ef65688810 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.328504] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6184f1ff-27ec-43b2-b2a3-f000b18f3d72 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.336422] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bb872cf-9ad8-4b43-a3f1-cfb615a9fa2a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.349867] env[62692]: DEBUG nova.compute.provider_tree [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.400596] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141408, 'name': PowerOffVM_Task, 'duration_secs': 0.349602} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.400910] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 936.401205] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 936.402174] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0def21be-e8cb-425b-a168-d31f31af9e24 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.409975] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 936.410287] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d2943dad-bcf2-43ca-9434-dbc483d4b03b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.449457] env[62692]: DEBUG nova.compute.manager [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 936.487123] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 936.487341] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 936.487614] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Deleting the datastore file [datastore1] 5a08fa8a-f9fe-4879-bb7b-baa04097df6c {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 936.488614] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8610273b-4072-49b3-88df-c4378b7d9bfd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.496279] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 936.496279] env[62692]: value = "task-1141410" [ 936.496279] env[62692]: _type = "Task" [ 936.496279] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.505528] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141410, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.651569] env[62692]: DEBUG oslo_concurrency.lockutils [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Acquiring lock "8a56e6df-eea1-41f4-9360-4f06d2f516a3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.651569] env[62692]: DEBUG oslo_concurrency.lockutils [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Lock "8a56e6df-eea1-41f4-9360-4f06d2f516a3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.651813] env[62692]: DEBUG oslo_concurrency.lockutils [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Acquiring lock "8a56e6df-eea1-41f4-9360-4f06d2f516a3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.652044] env[62692]: DEBUG oslo_concurrency.lockutils [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Lock "8a56e6df-eea1-41f4-9360-4f06d2f516a3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.652254] env[62692]: DEBUG oslo_concurrency.lockutils [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Lock "8a56e6df-eea1-41f4-9360-4f06d2f516a3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.654684] env[62692]: INFO nova.compute.manager [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Terminating instance [ 936.656644] env[62692]: DEBUG nova.compute.manager [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 936.656874] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 936.657814] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74789eed-9526-4157-a2a3-deff77292bba {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.666942] env[62692]: INFO nova.compute.manager [-] [instance: fac426b6-747a-455e-85a1-44e71b432180] Took 1.04 seconds to deallocate network for instance. [ 936.667319] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 936.668957] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-198d38f9-b9b9-4b32-83dc-b1aad0109fca {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.679789] env[62692]: DEBUG oslo_vmware.api [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Waiting for the task: (returnval){ [ 936.679789] env[62692]: value = "task-1141411" [ 936.679789] env[62692]: _type = "Task" [ 936.679789] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.690910] env[62692]: DEBUG oslo_vmware.api [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Task: {'id': task-1141411, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.740045] env[62692]: DEBUG oslo_vmware.api [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]522c02e0-2190-a4ab-fc51-270d1b500295, 'name': SearchDatastore_Task, 'duration_secs': 0.017282} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.745228] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Reconfiguring VM instance instance-00000035 to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 936.745544] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1e522ceb-8791-46cd-8909-f4832be3950e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.765293] env[62692]: DEBUG oslo_vmware.api [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Waiting for the task: (returnval){ [ 936.765293] env[62692]: value = "task-1141412" [ 936.765293] env[62692]: _type = "Task" [ 936.765293] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.770235] env[62692]: DEBUG nova.compute.manager [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 936.779217] env[62692]: DEBUG oslo_vmware.api [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141412, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.799401] env[62692]: DEBUG nova.virt.hardware [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 936.799688] env[62692]: DEBUG nova.virt.hardware [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 936.799846] env[62692]: DEBUG nova.virt.hardware [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 936.800048] env[62692]: DEBUG nova.virt.hardware [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 936.800210] env[62692]: DEBUG nova.virt.hardware [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 936.800365] env[62692]: DEBUG nova.virt.hardware [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 936.800583] env[62692]: DEBUG nova.virt.hardware [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 936.800751] env[62692]: DEBUG nova.virt.hardware [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 936.800926] env[62692]: DEBUG nova.virt.hardware [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 936.801110] env[62692]: DEBUG nova.virt.hardware [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 936.801291] env[62692]: DEBUG nova.virt.hardware [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 936.802184] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eeb4879-cb7e-47d0-9891-3a22372f52d9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.810131] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8efdc0e1-328d-494c-8c52-fdc6ce47e550 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.852481] env[62692]: DEBUG nova.scheduler.client.report [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 936.973628] env[62692]: DEBUG oslo_concurrency.lockutils [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.007175] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141410, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.29597} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.007476] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 937.007747] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 937.007959] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 937.175372] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.190809] env[62692]: DEBUG oslo_vmware.api [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Task: {'id': task-1141411, 'name': PowerOffVM_Task, 'duration_secs': 0.206129} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.191105] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 937.191334] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 937.191699] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cda4418d-de83-4473-b07c-6b5e41f0fa9b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.225931] env[62692]: DEBUG nova.compute.manager [req-538beae5-2068-4fc8-83a9-d6502fbdb369 req-1d53287d-767f-4107-859f-f4a7c9140a02 service nova] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Received event network-vif-plugged-62803691-31f7-422f-b373-a85721fe22b6 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 937.226196] env[62692]: DEBUG oslo_concurrency.lockutils [req-538beae5-2068-4fc8-83a9-d6502fbdb369 req-1d53287d-767f-4107-859f-f4a7c9140a02 service nova] Acquiring lock "e2b3b062-10e2-4a9c-b6b3-f0786de07b79-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.226417] env[62692]: DEBUG oslo_concurrency.lockutils [req-538beae5-2068-4fc8-83a9-d6502fbdb369 req-1d53287d-767f-4107-859f-f4a7c9140a02 service nova] Lock "e2b3b062-10e2-4a9c-b6b3-f0786de07b79-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.226629] env[62692]: DEBUG oslo_concurrency.lockutils [req-538beae5-2068-4fc8-83a9-d6502fbdb369 req-1d53287d-767f-4107-859f-f4a7c9140a02 service nova] Lock "e2b3b062-10e2-4a9c-b6b3-f0786de07b79-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.226891] env[62692]: DEBUG nova.compute.manager [req-538beae5-2068-4fc8-83a9-d6502fbdb369 req-1d53287d-767f-4107-859f-f4a7c9140a02 service nova] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] No waiting events found dispatching network-vif-plugged-62803691-31f7-422f-b373-a85721fe22b6 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 937.227297] env[62692]: WARNING nova.compute.manager [req-538beae5-2068-4fc8-83a9-d6502fbdb369 req-1d53287d-767f-4107-859f-f4a7c9140a02 service nova] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Received unexpected event network-vif-plugged-62803691-31f7-422f-b373-a85721fe22b6 for instance with vm_state building and task_state spawning. [ 937.276822] env[62692]: DEBUG oslo_vmware.api [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141412, 'name': ReconfigVM_Task, 'duration_secs': 0.319443} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.277189] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Reconfigured VM instance instance-00000035 to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 937.278111] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c61eef27-6f9e-4437-b0a5-164c56c827e2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.305223] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] 2ccb76cf-f641-4306-a137-fb2417285df9/2ccb76cf-f641-4306-a137-fb2417285df9.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 937.307078] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cb425f5e-0a8c-4de0-be88-f35efdf06b63 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.320544] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 937.320712] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 937.320902] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Deleting the datastore file [datastore2] 8a56e6df-eea1-41f4-9360-4f06d2f516a3 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 937.321230] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0297c071-babb-4781-8e10-1faf023f2588 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.330880] env[62692]: DEBUG oslo_vmware.api [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Waiting for the task: (returnval){ [ 937.330880] env[62692]: value = "task-1141414" [ 937.330880] env[62692]: _type = "Task" [ 937.330880] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.332459] env[62692]: DEBUG oslo_vmware.api [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Waiting for the task: (returnval){ [ 937.332459] env[62692]: value = "task-1141415" [ 937.332459] env[62692]: _type = "Task" [ 937.332459] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.353913] env[62692]: DEBUG oslo_vmware.api [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141415, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.354478] env[62692]: DEBUG oslo_vmware.api [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Task: {'id': task-1141414, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.355371] env[62692]: DEBUG nova.network.neutron [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Successfully updated port: 62803691-31f7-422f-b373-a85721fe22b6 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 937.357278] env[62692]: DEBUG oslo_concurrency.lockutils [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.613s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.358227] env[62692]: DEBUG nova.compute.manager [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 937.362659] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.919s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.362659] env[62692]: DEBUG nova.objects.instance [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lazy-loading 'resources' on Instance uuid ae44f074-a5c8-4259-99c6-9ce290b1570b {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 937.846260] env[62692]: DEBUG oslo_vmware.api [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141415, 'name': ReconfigVM_Task, 'duration_secs': 0.312551} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.849902] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Reconfigured VM instance instance-00000035 to attach disk [datastore1] 2ccb76cf-f641-4306-a137-fb2417285df9/2ccb76cf-f641-4306-a137-fb2417285df9.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 937.849902] env[62692]: DEBUG oslo_vmware.api [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Task: {'id': task-1141414, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177411} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.850608] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7c4768d-98fe-40ba-9b8d-4b4f4858b748 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.853144] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 937.853349] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 937.853537] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 937.853717] env[62692]: INFO nova.compute.manager [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Took 1.20 seconds to destroy the instance on the hypervisor. [ 937.853972] env[62692]: DEBUG oslo.service.loopingcall [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 937.854194] env[62692]: DEBUG nova.compute.manager [-] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 937.854292] env[62692]: DEBUG nova.network.neutron [-] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 937.879260] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Acquiring lock "refresh_cache-e2b3b062-10e2-4a9c-b6b3-f0786de07b79" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.879435] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Acquired lock "refresh_cache-e2b3b062-10e2-4a9c-b6b3-f0786de07b79" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.879588] env[62692]: DEBUG nova.network.neutron [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 937.884323] env[62692]: DEBUG nova.compute.utils [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 937.894974] env[62692]: DEBUG nova.compute.manager [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 937.894974] env[62692]: DEBUG nova.network.neutron [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 937.897218] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c031d97-7950-4085-b212-d406ab090d08 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.920219] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca648da-97be-4e22-a91d-3dca6668dff4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.942364] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee972c99-5164-48ab-b6e0-2e4b6c57f9da {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.955310] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 937.958357] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-71492375-3d21-4cf3-9134-2bfd76c21159 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.969472] env[62692]: DEBUG nova.network.neutron [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 937.969472] env[62692]: DEBUG oslo_vmware.api [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Waiting for the task: (returnval){ [ 937.969472] env[62692]: value = "task-1141416" [ 937.969472] env[62692]: _type = "Task" [ 937.969472] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.977658] env[62692]: DEBUG oslo_vmware.api [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141416, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.002190] env[62692]: DEBUG nova.policy [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b40c8976b544c059458aa4e0dbab160', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9cb8ae08824245a7911de7455e7fcb68', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 938.052734] env[62692]: DEBUG nova.virt.hardware [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 938.053054] env[62692]: DEBUG nova.virt.hardware [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 938.053262] env[62692]: DEBUG nova.virt.hardware [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 938.053504] env[62692]: DEBUG nova.virt.hardware [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 938.053674] env[62692]: DEBUG nova.virt.hardware [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 938.053863] env[62692]: DEBUG nova.virt.hardware [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 938.054151] env[62692]: DEBUG nova.virt.hardware [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 938.054357] env[62692]: DEBUG nova.virt.hardware [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 938.054569] env[62692]: DEBUG nova.virt.hardware [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 938.054746] env[62692]: DEBUG nova.virt.hardware [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 938.054926] env[62692]: DEBUG nova.virt.hardware [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 938.056327] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20da702c-f149-419c-bd6f-2f7923d89e34 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.069687] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4e8a463-7468-472f-a3de-745000d016ce {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.084636] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6c:33:e3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '678ebbe4-4c53-4eaf-a689-93981310f37d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6f540034-8dda-4def-adb2-0cdba3268510', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 938.092796] env[62692]: DEBUG oslo.service.loopingcall [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 938.095887] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 938.097200] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4bb2bf91-7dec-4c11-8c6f-8f2cee672b55 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.123755] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 938.123755] env[62692]: value = "task-1141417" [ 938.123755] env[62692]: _type = "Task" [ 938.123755] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.138930] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141417, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.154475] env[62692]: DEBUG nova.network.neutron [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Updating instance_info_cache with network_info: [{"id": "62803691-31f7-422f-b373-a85721fe22b6", "address": "fa:16:3e:81:4a:a0", "network": {"id": "5f0121e2-bbe3-4069-96a7-b16dfcf5514d", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-9909328-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54c19ee28b1c4b829ffa32a61a4d22a9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62803691-31", "ovs_interfaceid": "62803691-31f7-422f-b373-a85721fe22b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.393504] env[62692]: DEBUG nova.compute.manager [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 938.428207] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-013a13bb-db51-4f05-b190-1483cc547232 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.440023] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4431b407-74f5-4e50-ae90-835b60b11a42 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.478717] env[62692]: DEBUG nova.network.neutron [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Successfully created port: 2cb4b52c-0d89-4402-ae48-dc6f226c557c {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 938.484620] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d49bc2e-4ade-4512-b722-3e7d800e37ee {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.496403] env[62692]: DEBUG oslo_vmware.api [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141416, 'name': PowerOnVM_Task, 'duration_secs': 0.413746} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.496959] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 938.504487] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ce39c41-8205-4384-9f95-a5b67d61ce90 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.523180] env[62692]: DEBUG nova.compute.provider_tree [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 938.635916] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141417, 'name': CreateVM_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.661028] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Releasing lock "refresh_cache-e2b3b062-10e2-4a9c-b6b3-f0786de07b79" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.661028] env[62692]: DEBUG nova.compute.manager [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Instance network_info: |[{"id": "62803691-31f7-422f-b373-a85721fe22b6", "address": "fa:16:3e:81:4a:a0", "network": {"id": "5f0121e2-bbe3-4069-96a7-b16dfcf5514d", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-9909328-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54c19ee28b1c4b829ffa32a61a4d22a9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62803691-31", "ovs_interfaceid": "62803691-31f7-422f-b373-a85721fe22b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 938.661028] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:4a:a0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '93c5b7ce-4c84-40bc-884c-b2453e0eee69', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '62803691-31f7-422f-b373-a85721fe22b6', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 938.668408] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Creating folder: Project (54c19ee28b1c4b829ffa32a61a4d22a9). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 938.669420] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f8363dec-5bb9-45f5-abec-13ea052d450e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.685742] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Created folder: Project (54c19ee28b1c4b829ffa32a61a4d22a9) in parent group-v248868. [ 938.685742] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Creating folder: Instances. Parent ref: group-v248991. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 938.685742] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ea2f8e9d-044d-4d0f-9c49-5d3a42d70d69 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.699997] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Created folder: Instances in parent group-v248991. [ 938.700333] env[62692]: DEBUG oslo.service.loopingcall [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 938.700556] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 938.700780] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f679faa4-811f-48a6-9c6c-9bdd154b9fca {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.720442] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 938.720442] env[62692]: value = "task-1141420" [ 938.720442] env[62692]: _type = "Task" [ 938.720442] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.788731] env[62692]: DEBUG nova.network.neutron [-] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.028515] env[62692]: DEBUG nova.scheduler.client.report [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 939.136079] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141417, 'name': CreateVM_Task, 'duration_secs': 0.646071} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.136754] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 939.137022] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.137231] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.137575] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 939.137835] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e131330-b142-415a-bd90-ec4cfb225dc7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.142677] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 939.142677] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52caf191-17f7-19c0-bce1-3f6756a8000f" [ 939.142677] env[62692]: _type = "Task" [ 939.142677] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.150747] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52caf191-17f7-19c0-bce1-3f6756a8000f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.229312] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141420, 'name': CreateVM_Task, 'duration_secs': 0.457656} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.229488] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 939.230144] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.255607] env[62692]: DEBUG nova.compute.manager [req-a1b2f4b5-ddb4-4c3e-9b54-d7e01bb0f26f req-58fc0e14-9a65-42fe-828f-1e9a2ef14b72 service nova] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Received event network-changed-62803691-31f7-422f-b373-a85721fe22b6 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 939.255813] env[62692]: DEBUG nova.compute.manager [req-a1b2f4b5-ddb4-4c3e-9b54-d7e01bb0f26f req-58fc0e14-9a65-42fe-828f-1e9a2ef14b72 service nova] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Refreshing instance network info cache due to event network-changed-62803691-31f7-422f-b373-a85721fe22b6. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 939.256044] env[62692]: DEBUG oslo_concurrency.lockutils [req-a1b2f4b5-ddb4-4c3e-9b54-d7e01bb0f26f req-58fc0e14-9a65-42fe-828f-1e9a2ef14b72 service nova] Acquiring lock "refresh_cache-e2b3b062-10e2-4a9c-b6b3-f0786de07b79" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.256198] env[62692]: DEBUG oslo_concurrency.lockutils [req-a1b2f4b5-ddb4-4c3e-9b54-d7e01bb0f26f req-58fc0e14-9a65-42fe-828f-1e9a2ef14b72 service nova] Acquired lock "refresh_cache-e2b3b062-10e2-4a9c-b6b3-f0786de07b79" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.256367] env[62692]: DEBUG nova.network.neutron [req-a1b2f4b5-ddb4-4c3e-9b54-d7e01bb0f26f req-58fc0e14-9a65-42fe-828f-1e9a2ef14b72 service nova] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Refreshing network info cache for port 62803691-31f7-422f-b373-a85721fe22b6 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 939.290804] env[62692]: INFO nova.compute.manager [-] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Took 1.44 seconds to deallocate network for instance. [ 939.404510] env[62692]: DEBUG nova.compute.manager [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 939.429008] env[62692]: DEBUG nova.virt.hardware [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 939.429331] env[62692]: DEBUG nova.virt.hardware [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 939.429504] env[62692]: DEBUG nova.virt.hardware [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 939.429694] env[62692]: DEBUG nova.virt.hardware [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 939.429842] env[62692]: DEBUG nova.virt.hardware [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 939.429994] env[62692]: DEBUG nova.virt.hardware [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 939.430218] env[62692]: DEBUG nova.virt.hardware [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 939.430500] env[62692]: DEBUG nova.virt.hardware [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 939.430679] env[62692]: DEBUG nova.virt.hardware [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 939.430848] env[62692]: DEBUG nova.virt.hardware [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 939.431032] env[62692]: DEBUG nova.virt.hardware [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 939.431871] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8e941e8-5d17-401b-9031-7a0278174632 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.439960] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad974f6b-2cd0-4175-a573-fc57650b2fab {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.516427] env[62692]: INFO nova.compute.manager [None req-133db2e5-0094-4474-8285-9ac9104eda21 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Updating instance to original state: 'active' [ 939.533991] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.172s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.536412] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.204s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.538103] env[62692]: INFO nova.compute.claims [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 939.556501] env[62692]: INFO nova.scheduler.client.report [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Deleted allocations for instance ae44f074-a5c8-4259-99c6-9ce290b1570b [ 939.655367] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52caf191-17f7-19c0-bce1-3f6756a8000f, 'name': SearchDatastore_Task, 'duration_secs': 0.027183} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.655782] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.656144] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 939.656410] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.656599] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.656864] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 939.657247] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.657638] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 939.657925] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1be34294-6d25-48cd-9857-c0819f4453bd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.660059] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01349a4e-36e3-4a94-a980-09b0c48516da {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.665580] env[62692]: DEBUG oslo_vmware.api [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Waiting for the task: (returnval){ [ 939.665580] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]525bfea0-61e2-8225-951e-830c55c76e98" [ 939.665580] env[62692]: _type = "Task" [ 939.665580] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.675521] env[62692]: DEBUG oslo_vmware.api [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]525bfea0-61e2-8225-951e-830c55c76e98, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.677975] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 939.678221] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 939.678866] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0337859e-40e3-41b7-a073-a0bbcd910e30 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.684332] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 939.684332] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52d5e3b0-886d-6563-14ad-aefb07eb6867" [ 939.684332] env[62692]: _type = "Task" [ 939.684332] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.691861] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52d5e3b0-886d-6563-14ad-aefb07eb6867, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.797025] env[62692]: DEBUG oslo_concurrency.lockutils [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.064411] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3dc69718-db1e-4577-897b-e883aab647a0 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lock "ae44f074-a5c8-4259-99c6-9ce290b1570b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.653s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.129168] env[62692]: DEBUG nova.network.neutron [req-a1b2f4b5-ddb4-4c3e-9b54-d7e01bb0f26f req-58fc0e14-9a65-42fe-828f-1e9a2ef14b72 service nova] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Updated VIF entry in instance network info cache for port 62803691-31f7-422f-b373-a85721fe22b6. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 940.132986] env[62692]: DEBUG nova.network.neutron [req-a1b2f4b5-ddb4-4c3e-9b54-d7e01bb0f26f req-58fc0e14-9a65-42fe-828f-1e9a2ef14b72 service nova] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Updating instance_info_cache with network_info: [{"id": "62803691-31f7-422f-b373-a85721fe22b6", "address": "fa:16:3e:81:4a:a0", "network": {"id": "5f0121e2-bbe3-4069-96a7-b16dfcf5514d", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-9909328-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54c19ee28b1c4b829ffa32a61a4d22a9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62803691-31", "ovs_interfaceid": "62803691-31f7-422f-b373-a85721fe22b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.175463] env[62692]: DEBUG oslo_vmware.api [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]525bfea0-61e2-8225-951e-830c55c76e98, 'name': SearchDatastore_Task, 'duration_secs': 0.011639} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.175779] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.176025] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 940.176246] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.193480] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52d5e3b0-886d-6563-14ad-aefb07eb6867, 'name': SearchDatastore_Task, 'duration_secs': 0.017781} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.194289] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-103130df-042a-4fa0-a891-df3bccaab60f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.199668] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 940.199668] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]520f637c-8068-ee7a-9dfd-7ed4586e2b47" [ 940.199668] env[62692]: _type = "Task" [ 940.199668] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.209928] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]520f637c-8068-ee7a-9dfd-7ed4586e2b47, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.286430] env[62692]: DEBUG nova.network.neutron [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Successfully updated port: 2cb4b52c-0d89-4402-ae48-dc6f226c557c {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 940.636438] env[62692]: DEBUG oslo_concurrency.lockutils [req-a1b2f4b5-ddb4-4c3e-9b54-d7e01bb0f26f req-58fc0e14-9a65-42fe-828f-1e9a2ef14b72 service nova] Releasing lock "refresh_cache-e2b3b062-10e2-4a9c-b6b3-f0786de07b79" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.636738] env[62692]: DEBUG nova.compute.manager [req-a1b2f4b5-ddb4-4c3e-9b54-d7e01bb0f26f req-58fc0e14-9a65-42fe-828f-1e9a2ef14b72 service nova] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Received event network-vif-deleted-1ff8b50f-db1f-4a27-8ce4-228ab7446c64 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 940.713511] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]520f637c-8068-ee7a-9dfd-7ed4586e2b47, 'name': SearchDatastore_Task, 'duration_secs': 0.012178} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.713828] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.714066] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 5a08fa8a-f9fe-4879-bb7b-baa04097df6c/5a08fa8a-f9fe-4879-bb7b-baa04097df6c.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 940.714360] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.714652] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 940.715530] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f1f51f3e-4c4f-49be-a9c2-70b9d5e161f1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.720029] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-84887846-e326-4724-8cfa-39c22138187b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.726836] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 940.726836] env[62692]: value = "task-1141422" [ 940.726836] env[62692]: _type = "Task" [ 940.726836] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.730963] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 940.731179] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 940.734601] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ff61890-a0d7-42fc-ba10-aff5cd4498c9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.740082] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141422, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.743593] env[62692]: DEBUG oslo_vmware.api [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Waiting for the task: (returnval){ [ 940.743593] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]529f69ce-8c5e-d1e8-e765-1caeb86310a4" [ 940.743593] env[62692]: _type = "Task" [ 940.743593] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.757580] env[62692]: DEBUG oslo_vmware.api [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]529f69ce-8c5e-d1e8-e765-1caeb86310a4, 'name': SearchDatastore_Task, 'duration_secs': 0.009491} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.758403] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e18af847-5424-4b5d-ae4e-c55883966e3d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.763338] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquiring lock "2ccb76cf-f641-4306-a137-fb2417285df9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.763577] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Lock "2ccb76cf-f641-4306-a137-fb2417285df9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.763782] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquiring lock "2ccb76cf-f641-4306-a137-fb2417285df9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.763970] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Lock "2ccb76cf-f641-4306-a137-fb2417285df9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.764159] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Lock "2ccb76cf-f641-4306-a137-fb2417285df9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.766531] env[62692]: INFO nova.compute.manager [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Terminating instance [ 940.769018] env[62692]: DEBUG oslo_vmware.api [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Waiting for the task: (returnval){ [ 940.769018] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5280dbc3-4cfe-af60-1587-74b348e10a5d" [ 940.769018] env[62692]: _type = "Task" [ 940.769018] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.769620] env[62692]: DEBUG nova.compute.manager [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 940.769845] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 940.770769] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8beb45af-dc39-464f-8078-03c8e3bfe583 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.784459] env[62692]: DEBUG oslo_vmware.api [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5280dbc3-4cfe-af60-1587-74b348e10a5d, 'name': SearchDatastore_Task, 'duration_secs': 0.009487} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.786571] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.786845] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] e2b3b062-10e2-4a9c-b6b3-f0786de07b79/e2b3b062-10e2-4a9c-b6b3-f0786de07b79.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 940.787214] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 940.787413] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4cca634f-6916-45a3-b2a8-3e2394a5bcae {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.789265] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5e5ae954-47f2-482f-9f91-674bd3b72248 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.791122] env[62692]: DEBUG oslo_concurrency.lockutils [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "refresh_cache-3f6c9744-a6e3-43f3-8b6b-624ba178747d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.791260] env[62692]: DEBUG oslo_concurrency.lockutils [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired lock "refresh_cache-3f6c9744-a6e3-43f3-8b6b-624ba178747d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.791419] env[62692]: DEBUG nova.network.neutron [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 940.803951] env[62692]: DEBUG oslo_vmware.api [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Waiting for the task: (returnval){ [ 940.803951] env[62692]: value = "task-1141423" [ 940.803951] env[62692]: _type = "Task" [ 940.803951] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.804231] env[62692]: DEBUG oslo_vmware.api [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Waiting for the task: (returnval){ [ 940.804231] env[62692]: value = "task-1141424" [ 940.804231] env[62692]: _type = "Task" [ 940.804231] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.823622] env[62692]: DEBUG oslo_vmware.api [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Task: {'id': task-1141424, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.823622] env[62692]: DEBUG oslo_vmware.api [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141423, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.979683] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb565b76-d34f-452b-b970-28c95b224533 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.988915] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eab90f8e-b875-4290-859a-15ff9062e47b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.023100] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4350871e-5cfe-4d56-bf1c-fa2d9e848132 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.031581] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47bcd06f-7422-4d96-b7ef-732f9db4123e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.045730] env[62692]: DEBUG nova.compute.provider_tree [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.238649] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141422, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.284888] env[62692]: DEBUG nova.compute.manager [req-9c10fa36-a96e-4c0a-a34e-17dacd28f0c2 req-c784b149-dc49-4dfb-85d9-bd5fbe25efae service nova] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Received event network-vif-plugged-2cb4b52c-0d89-4402-ae48-dc6f226c557c {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 941.285138] env[62692]: DEBUG oslo_concurrency.lockutils [req-9c10fa36-a96e-4c0a-a34e-17dacd28f0c2 req-c784b149-dc49-4dfb-85d9-bd5fbe25efae service nova] Acquiring lock "3f6c9744-a6e3-43f3-8b6b-624ba178747d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.286027] env[62692]: DEBUG oslo_concurrency.lockutils [req-9c10fa36-a96e-4c0a-a34e-17dacd28f0c2 req-c784b149-dc49-4dfb-85d9-bd5fbe25efae service nova] Lock "3f6c9744-a6e3-43f3-8b6b-624ba178747d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.286027] env[62692]: DEBUG oslo_concurrency.lockutils [req-9c10fa36-a96e-4c0a-a34e-17dacd28f0c2 req-c784b149-dc49-4dfb-85d9-bd5fbe25efae service nova] Lock "3f6c9744-a6e3-43f3-8b6b-624ba178747d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.286027] env[62692]: DEBUG nova.compute.manager [req-9c10fa36-a96e-4c0a-a34e-17dacd28f0c2 req-c784b149-dc49-4dfb-85d9-bd5fbe25efae service nova] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] No waiting events found dispatching network-vif-plugged-2cb4b52c-0d89-4402-ae48-dc6f226c557c {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 941.286027] env[62692]: WARNING nova.compute.manager [req-9c10fa36-a96e-4c0a-a34e-17dacd28f0c2 req-c784b149-dc49-4dfb-85d9-bd5fbe25efae service nova] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Received unexpected event network-vif-plugged-2cb4b52c-0d89-4402-ae48-dc6f226c557c for instance with vm_state building and task_state spawning. [ 941.286307] env[62692]: DEBUG nova.compute.manager [req-9c10fa36-a96e-4c0a-a34e-17dacd28f0c2 req-c784b149-dc49-4dfb-85d9-bd5fbe25efae service nova] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Received event network-changed-2cb4b52c-0d89-4402-ae48-dc6f226c557c {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 941.286307] env[62692]: DEBUG nova.compute.manager [req-9c10fa36-a96e-4c0a-a34e-17dacd28f0c2 req-c784b149-dc49-4dfb-85d9-bd5fbe25efae service nova] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Refreshing instance network info cache due to event network-changed-2cb4b52c-0d89-4402-ae48-dc6f226c557c. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 941.286436] env[62692]: DEBUG oslo_concurrency.lockutils [req-9c10fa36-a96e-4c0a-a34e-17dacd28f0c2 req-c784b149-dc49-4dfb-85d9-bd5fbe25efae service nova] Acquiring lock "refresh_cache-3f6c9744-a6e3-43f3-8b6b-624ba178747d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.318044] env[62692]: DEBUG oslo_vmware.api [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141423, 'name': PowerOffVM_Task, 'duration_secs': 0.215317} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.321184] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 941.321428] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 941.322134] env[62692]: DEBUG oslo_vmware.api [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Task: {'id': task-1141424, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.322378] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6a042a2f-29de-43eb-8b05-63a7d101cfeb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.329410] env[62692]: DEBUG nova.network.neutron [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 941.498378] env[62692]: DEBUG nova.network.neutron [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Updating instance_info_cache with network_info: [{"id": "2cb4b52c-0d89-4402-ae48-dc6f226c557c", "address": "fa:16:3e:c8:b1:41", "network": {"id": "f6352999-f943-431d-b54a-c9dc1d810a85", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-121037111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9cb8ae08824245a7911de7455e7fcb68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c842425c-544e-4ce2-9657-512723bd318e", "external-id": "nsx-vlan-transportzone-80", "segmentation_id": 80, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cb4b52c-0d", "ovs_interfaceid": "2cb4b52c-0d89-4402-ae48-dc6f226c557c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.550023] env[62692]: DEBUG nova.scheduler.client.report [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 941.736900] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141422, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.646163} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.737212] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 5a08fa8a-f9fe-4879-bb7b-baa04097df6c/5a08fa8a-f9fe-4879-bb7b-baa04097df6c.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 941.737436] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 941.737689] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bab86ab2-d9aa-4d31-a9ce-ac56de277779 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.743669] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 941.743669] env[62692]: value = "task-1141426" [ 941.743669] env[62692]: _type = "Task" [ 941.743669] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.751025] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141426, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.815759] env[62692]: DEBUG oslo_vmware.api [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Task: {'id': task-1141424, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.890719} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.816038] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] e2b3b062-10e2-4a9c-b6b3-f0786de07b79/e2b3b062-10e2-4a9c-b6b3-f0786de07b79.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 941.816260] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 941.816512] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1ed1f24a-297e-40f2-b072-9a1596aa45c0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.823356] env[62692]: DEBUG oslo_vmware.api [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Waiting for the task: (returnval){ [ 941.823356] env[62692]: value = "task-1141427" [ 941.823356] env[62692]: _type = "Task" [ 941.823356] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.831478] env[62692]: DEBUG oslo_vmware.api [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Task: {'id': task-1141427, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.001837] env[62692]: DEBUG oslo_concurrency.lockutils [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Releasing lock "refresh_cache-3f6c9744-a6e3-43f3-8b6b-624ba178747d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.001837] env[62692]: DEBUG nova.compute.manager [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Instance network_info: |[{"id": "2cb4b52c-0d89-4402-ae48-dc6f226c557c", "address": "fa:16:3e:c8:b1:41", "network": {"id": "f6352999-f943-431d-b54a-c9dc1d810a85", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-121037111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9cb8ae08824245a7911de7455e7fcb68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c842425c-544e-4ce2-9657-512723bd318e", "external-id": "nsx-vlan-transportzone-80", "segmentation_id": 80, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cb4b52c-0d", "ovs_interfaceid": "2cb4b52c-0d89-4402-ae48-dc6f226c557c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 942.002042] env[62692]: DEBUG oslo_concurrency.lockutils [req-9c10fa36-a96e-4c0a-a34e-17dacd28f0c2 req-c784b149-dc49-4dfb-85d9-bd5fbe25efae service nova] Acquired lock "refresh_cache-3f6c9744-a6e3-43f3-8b6b-624ba178747d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.002279] env[62692]: DEBUG nova.network.neutron [req-9c10fa36-a96e-4c0a-a34e-17dacd28f0c2 req-c784b149-dc49-4dfb-85d9-bd5fbe25efae service nova] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Refreshing network info cache for port 2cb4b52c-0d89-4402-ae48-dc6f226c557c {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 942.003481] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c8:b1:41', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c842425c-544e-4ce2-9657-512723bd318e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2cb4b52c-0d89-4402-ae48-dc6f226c557c', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 942.012929] env[62692]: DEBUG oslo.service.loopingcall [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 942.016135] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 942.016708] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-151e5f6f-7306-4401-ad3d-7a2e62e70502 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.037549] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 942.037549] env[62692]: value = "task-1141428" [ 942.037549] env[62692]: _type = "Task" [ 942.037549] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.045844] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141428, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.054991] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.518s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.055704] env[62692]: DEBUG nova.compute.manager [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 942.058677] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.947s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.058952] env[62692]: DEBUG nova.objects.instance [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Lazy-loading 'resources' on Instance uuid 628cc505-3edf-4066-91be-da009ebcf219 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 942.253490] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141426, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.316839] env[62692]: DEBUG nova.network.neutron [req-9c10fa36-a96e-4c0a-a34e-17dacd28f0c2 req-c784b149-dc49-4dfb-85d9-bd5fbe25efae service nova] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Updated VIF entry in instance network info cache for port 2cb4b52c-0d89-4402-ae48-dc6f226c557c. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 942.317251] env[62692]: DEBUG nova.network.neutron [req-9c10fa36-a96e-4c0a-a34e-17dacd28f0c2 req-c784b149-dc49-4dfb-85d9-bd5fbe25efae service nova] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Updating instance_info_cache with network_info: [{"id": "2cb4b52c-0d89-4402-ae48-dc6f226c557c", "address": "fa:16:3e:c8:b1:41", "network": {"id": "f6352999-f943-431d-b54a-c9dc1d810a85", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-121037111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9cb8ae08824245a7911de7455e7fcb68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c842425c-544e-4ce2-9657-512723bd318e", "external-id": "nsx-vlan-transportzone-80", "segmentation_id": 80, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cb4b52c-0d", "ovs_interfaceid": "2cb4b52c-0d89-4402-ae48-dc6f226c557c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.332852] env[62692]: DEBUG oslo_vmware.api [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Task: {'id': task-1141427, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076011} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.333766] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 942.334571] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35411dec-d6a5-4e19-bef6-fccdff7de036 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.359527] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] e2b3b062-10e2-4a9c-b6b3-f0786de07b79/e2b3b062-10e2-4a9c-b6b3-f0786de07b79.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 942.360155] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fad709ad-73b4-4442-94bd-8c85b09818c3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.380473] env[62692]: DEBUG oslo_vmware.api [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Waiting for the task: (returnval){ [ 942.380473] env[62692]: value = "task-1141429" [ 942.380473] env[62692]: _type = "Task" [ 942.380473] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.388679] env[62692]: DEBUG oslo_vmware.api [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Task: {'id': task-1141429, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.463130] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 942.464032] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 942.464032] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Deleting the datastore file [datastore1] 2ccb76cf-f641-4306-a137-fb2417285df9 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 942.464267] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8fbcad0d-126d-4318-ac0b-452ceb6180a6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.470624] env[62692]: DEBUG oslo_vmware.api [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Waiting for the task: (returnval){ [ 942.470624] env[62692]: value = "task-1141430" [ 942.470624] env[62692]: _type = "Task" [ 942.470624] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.479079] env[62692]: DEBUG oslo_vmware.api [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141430, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.548533] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141428, 'name': CreateVM_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.564176] env[62692]: DEBUG nova.compute.utils [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 942.568723] env[62692]: DEBUG nova.compute.manager [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 942.568946] env[62692]: DEBUG nova.network.neutron [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 942.571219] env[62692]: DEBUG oslo_concurrency.lockutils [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Acquiring lock "abee88ff-5cf5-4bf1-91e0-93b19cf30046" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.571429] env[62692]: DEBUG oslo_concurrency.lockutils [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lock "abee88ff-5cf5-4bf1-91e0-93b19cf30046" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.572045] env[62692]: DEBUG oslo_concurrency.lockutils [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Acquiring lock "abee88ff-5cf5-4bf1-91e0-93b19cf30046-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.572045] env[62692]: DEBUG oslo_concurrency.lockutils [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lock "abee88ff-5cf5-4bf1-91e0-93b19cf30046-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.572204] env[62692]: DEBUG oslo_concurrency.lockutils [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lock "abee88ff-5cf5-4bf1-91e0-93b19cf30046-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.575606] env[62692]: INFO nova.compute.manager [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Terminating instance [ 942.579017] env[62692]: DEBUG nova.compute.manager [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 942.579017] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 942.579017] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-373a4910-fe74-4a22-950f-d31ac28f1617 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.586968] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 942.587270] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e1f9ca96-a89a-4fae-b726-b15cc0eb108e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.597471] env[62692]: DEBUG oslo_vmware.api [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 942.597471] env[62692]: value = "task-1141431" [ 942.597471] env[62692]: _type = "Task" [ 942.597471] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.606440] env[62692]: DEBUG oslo_vmware.api [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1141431, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.638171] env[62692]: DEBUG nova.policy [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '016134ac36fd4bd893226ae95a955e28', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37921f67664544478cb2568def068b11', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 942.757304] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141426, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.823765] env[62692]: DEBUG oslo_concurrency.lockutils [req-9c10fa36-a96e-4c0a-a34e-17dacd28f0c2 req-c784b149-dc49-4dfb-85d9-bd5fbe25efae service nova] Releasing lock "refresh_cache-3f6c9744-a6e3-43f3-8b6b-624ba178747d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.892451] env[62692]: DEBUG oslo_vmware.api [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Task: {'id': task-1141429, 'name': ReconfigVM_Task, 'duration_secs': 0.40177} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.895304] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Reconfigured VM instance instance-0000004a to attach disk [datastore1] e2b3b062-10e2-4a9c-b6b3-f0786de07b79/e2b3b062-10e2-4a9c-b6b3-f0786de07b79.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 942.896306] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2f5a87f3-a909-4cc3-9da9-f89c2430f028 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.906453] env[62692]: DEBUG oslo_vmware.api [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Waiting for the task: (returnval){ [ 942.906453] env[62692]: value = "task-1141432" [ 942.906453] env[62692]: _type = "Task" [ 942.906453] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.924828] env[62692]: DEBUG oslo_vmware.api [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Task: {'id': task-1141432, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.981794] env[62692]: DEBUG oslo_vmware.api [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Task: {'id': task-1141430, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.438949} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.984926] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 942.985274] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 942.985799] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 942.985799] env[62692]: INFO nova.compute.manager [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Took 2.22 seconds to destroy the instance on the hypervisor. [ 942.986062] env[62692]: DEBUG oslo.service.loopingcall [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 942.986509] env[62692]: DEBUG nova.compute.manager [-] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 942.986621] env[62692]: DEBUG nova.network.neutron [-] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 943.014291] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abd96f95-dcc5-4941-9ea2-1219a9216ae6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.023801] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d7e752-78b4-4c4b-8af6-b4ee5eade3ac {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.030444] env[62692]: DEBUG nova.network.neutron [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Successfully created port: f1187f72-32ea-484c-9bde-f40c9a5a3a03 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 943.066200] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5bd33bc-c177-45cd-8e26-cbc89b0271b4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.073764] env[62692]: DEBUG nova.compute.manager [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 943.082164] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141428, 'name': CreateVM_Task, 'duration_secs': 0.913667} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.084487] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 943.085369] env[62692]: DEBUG oslo_concurrency.lockutils [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.085535] env[62692]: DEBUG oslo_concurrency.lockutils [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.085863] env[62692]: DEBUG oslo_concurrency.lockutils [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 943.088183] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40944ec8-6a2e-41ba-8e49-bf7863942931 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.092763] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2449daf4-166e-4967-989b-d7eb12cb262f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.111746] env[62692]: DEBUG nova.compute.provider_tree [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.113163] env[62692]: DEBUG oslo_vmware.api [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 943.113163] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5223d125-766c-2b3a-e045-70a62c078c04" [ 943.113163] env[62692]: _type = "Task" [ 943.113163] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.121106] env[62692]: DEBUG oslo_vmware.api [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1141431, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.127652] env[62692]: DEBUG oslo_vmware.api [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5223d125-766c-2b3a-e045-70a62c078c04, 'name': SearchDatastore_Task, 'duration_secs': 0.010638} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.127933] env[62692]: DEBUG oslo_concurrency.lockutils [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.128190] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 943.128424] env[62692]: DEBUG oslo_concurrency.lockutils [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.128574] env[62692]: DEBUG oslo_concurrency.lockutils [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.128754] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 943.129013] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f221e5a8-3039-4dd4-8391-a73cf7fb3929 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.137082] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 943.137273] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 943.137974] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0fcc5925-1337-4430-849b-64169979fbab {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.143460] env[62692]: DEBUG oslo_vmware.api [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 943.143460] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52aeea2a-61f0-1e88-83a5-d0fa94803541" [ 943.143460] env[62692]: _type = "Task" [ 943.143460] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.151446] env[62692]: DEBUG oslo_vmware.api [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52aeea2a-61f0-1e88-83a5-d0fa94803541, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.257459] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141426, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.418751] env[62692]: DEBUG oslo_vmware.api [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Task: {'id': task-1141432, 'name': Rename_Task, 'duration_secs': 0.190274} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.419046] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 943.419366] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-65a98315-8e54-4dc6-ab4e-d42e2b3cf2c0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.427949] env[62692]: DEBUG oslo_vmware.api [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Waiting for the task: (returnval){ [ 943.427949] env[62692]: value = "task-1141433" [ 943.427949] env[62692]: _type = "Task" [ 943.427949] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.438662] env[62692]: DEBUG oslo_vmware.api [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Task: {'id': task-1141433, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.612865] env[62692]: DEBUG oslo_vmware.api [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1141431, 'name': PowerOffVM_Task, 'duration_secs': 0.654698} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.612865] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 943.612865] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 943.613073] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-98540a09-c5d3-4d30-a8fa-a1bc973c6388 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.615645] env[62692]: DEBUG nova.scheduler.client.report [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 943.662069] env[62692]: DEBUG oslo_vmware.api [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52aeea2a-61f0-1e88-83a5-d0fa94803541, 'name': SearchDatastore_Task, 'duration_secs': 0.010102} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.663052] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-800cdbfa-d487-4c2a-9d8b-5ed1fd1062de {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.670272] env[62692]: DEBUG oslo_vmware.api [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 943.670272] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5212b44b-7a78-1094-fc42-3e7d658fa930" [ 943.670272] env[62692]: _type = "Task" [ 943.670272] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.681531] env[62692]: DEBUG oslo_vmware.api [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5212b44b-7a78-1094-fc42-3e7d658fa930, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.683849] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 943.683849] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 943.683849] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Deleting the datastore file [datastore1] abee88ff-5cf5-4bf1-91e0-93b19cf30046 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 943.684130] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d6520674-1191-431b-87fb-1b97ddd18bcd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.693551] env[62692]: DEBUG oslo_vmware.api [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for the task: (returnval){ [ 943.693551] env[62692]: value = "task-1141435" [ 943.693551] env[62692]: _type = "Task" [ 943.693551] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.702878] env[62692]: DEBUG oslo_vmware.api [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1141435, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.756901] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141426, 'name': ExtendVirtualDisk_Task, 'duration_secs': 1.550701} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.757236] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 943.758071] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3214fdce-9445-4e4c-8261-fa567ac79c00 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.783828] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Reconfiguring VM instance instance-00000026 to attach disk [datastore1] 5a08fa8a-f9fe-4879-bb7b-baa04097df6c/5a08fa8a-f9fe-4879-bb7b-baa04097df6c.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 943.784556] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-791d615f-45c4-4500-84f6-5ca1bbd027e6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.801804] env[62692]: DEBUG nova.compute.manager [req-1239e12b-3802-4e68-8037-a8ce5e37a5f0 req-986a7fed-0c64-4ada-88a0-70ff077077c8 service nova] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Received event network-vif-deleted-951b0add-cfaa-4506-a6ac-56bcb8d1041e {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 943.802064] env[62692]: INFO nova.compute.manager [req-1239e12b-3802-4e68-8037-a8ce5e37a5f0 req-986a7fed-0c64-4ada-88a0-70ff077077c8 service nova] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Neutron deleted interface 951b0add-cfaa-4506-a6ac-56bcb8d1041e; detaching it from the instance and deleting it from the info cache [ 943.802397] env[62692]: DEBUG nova.network.neutron [req-1239e12b-3802-4e68-8037-a8ce5e37a5f0 req-986a7fed-0c64-4ada-88a0-70ff077077c8 service nova] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.810070] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 943.810070] env[62692]: value = "task-1141436" [ 943.810070] env[62692]: _type = "Task" [ 943.810070] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.820883] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141436, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.939799] env[62692]: DEBUG oslo_vmware.api [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Task: {'id': task-1141433, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.086950] env[62692]: DEBUG nova.compute.manager [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 944.112943] env[62692]: DEBUG nova.virt.hardware [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 944.113235] env[62692]: DEBUG nova.virt.hardware [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 944.113408] env[62692]: DEBUG nova.virt.hardware [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 944.113588] env[62692]: DEBUG nova.virt.hardware [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 944.113758] env[62692]: DEBUG nova.virt.hardware [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 944.113911] env[62692]: DEBUG nova.virt.hardware [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 944.114140] env[62692]: DEBUG nova.virt.hardware [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 944.114308] env[62692]: DEBUG nova.virt.hardware [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 944.114483] env[62692]: DEBUG nova.virt.hardware [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 944.114650] env[62692]: DEBUG nova.virt.hardware [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 944.114827] env[62692]: DEBUG nova.virt.hardware [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 944.115698] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba45e582-19b5-44f3-bc27-52f9d55b4446 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.121375] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.063s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.126189] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.864s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.127758] env[62692]: INFO nova.compute.claims [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 944.130473] env[62692]: DEBUG nova.network.neutron [-] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.132814] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82803019-f594-49e1-b384-7dd1115e4815 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.152658] env[62692]: INFO nova.scheduler.client.report [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Deleted allocations for instance 628cc505-3edf-4066-91be-da009ebcf219 [ 944.181988] env[62692]: DEBUG oslo_vmware.api [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5212b44b-7a78-1094-fc42-3e7d658fa930, 'name': SearchDatastore_Task, 'duration_secs': 0.015238} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.182289] env[62692]: DEBUG oslo_concurrency.lockutils [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.182555] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 3f6c9744-a6e3-43f3-8b6b-624ba178747d/3f6c9744-a6e3-43f3-8b6b-624ba178747d.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 944.182816] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-73a1403c-eee6-4461-adb6-dcadf4d52bf8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.190527] env[62692]: DEBUG oslo_vmware.api [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 944.190527] env[62692]: value = "task-1141437" [ 944.190527] env[62692]: _type = "Task" [ 944.190527] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.202393] env[62692]: DEBUG oslo_vmware.api [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141437, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.205437] env[62692]: DEBUG oslo_vmware.api [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Task: {'id': task-1141435, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.217186} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.206317] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 944.206523] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 944.206713] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 944.206890] env[62692]: INFO nova.compute.manager [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Took 1.63 seconds to destroy the instance on the hypervisor. [ 944.207203] env[62692]: DEBUG oslo.service.loopingcall [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 944.207376] env[62692]: DEBUG nova.compute.manager [-] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 944.207474] env[62692]: DEBUG nova.network.neutron [-] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 944.305743] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-34bb63a5-5e16-4b76-b236-b7ff37d57a07 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.320262] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af2d0da5-d3fc-4274-a7e2-d4aa759b3ed8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.334667] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141436, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.356679] env[62692]: DEBUG nova.compute.manager [req-1239e12b-3802-4e68-8037-a8ce5e37a5f0 req-986a7fed-0c64-4ada-88a0-70ff077077c8 service nova] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Detach interface failed, port_id=951b0add-cfaa-4506-a6ac-56bcb8d1041e, reason: Instance 2ccb76cf-f641-4306-a137-fb2417285df9 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 944.440541] env[62692]: DEBUG oslo_vmware.api [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Task: {'id': task-1141433, 'name': PowerOnVM_Task, 'duration_secs': 0.603392} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.440842] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 944.441080] env[62692]: INFO nova.compute.manager [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Took 7.67 seconds to spawn the instance on the hypervisor. [ 944.441279] env[62692]: DEBUG nova.compute.manager [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 944.442146] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b934afe-8a62-4dfd-be24-c1a6a2fbb646 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.638793] env[62692]: INFO nova.compute.manager [-] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Took 1.65 seconds to deallocate network for instance. [ 944.662799] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1cec2b8f-7d38-400c-b98e-3d85c2a57d7f tempest-SecurityGroupsTestJSON-144054308 tempest-SecurityGroupsTestJSON-144054308-project-member] Lock "628cc505-3edf-4066-91be-da009ebcf219" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.514s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.709154] env[62692]: DEBUG oslo_vmware.api [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141437, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.823223] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141436, 'name': ReconfigVM_Task, 'duration_secs': 0.52641} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.823661] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Reconfigured VM instance instance-00000026 to attach disk [datastore1] 5a08fa8a-f9fe-4879-bb7b-baa04097df6c/5a08fa8a-f9fe-4879-bb7b-baa04097df6c.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 944.824345] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0cadab29-d220-4b67-9fb0-fcf47536f24c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.832321] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 944.832321] env[62692]: value = "task-1141438" [ 944.832321] env[62692]: _type = "Task" [ 944.832321] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.841071] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141438, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.960901] env[62692]: INFO nova.compute.manager [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Took 34.62 seconds to build instance. [ 945.100374] env[62692]: DEBUG nova.network.neutron [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Successfully updated port: f1187f72-32ea-484c-9bde-f40c9a5a3a03 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 945.149384] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.205963] env[62692]: DEBUG oslo_vmware.api [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141437, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.553879} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.206250] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 3f6c9744-a6e3-43f3-8b6b-624ba178747d/3f6c9744-a6e3-43f3-8b6b-624ba178747d.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 945.206459] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 945.206963] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-62c8d8b9-1d9b-4c75-b347-1b508f1419fe {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.213764] env[62692]: DEBUG oslo_vmware.api [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 945.213764] env[62692]: value = "task-1141439" [ 945.213764] env[62692]: _type = "Task" [ 945.213764] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.226030] env[62692]: DEBUG oslo_vmware.api [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141439, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.249149] env[62692]: DEBUG nova.network.neutron [-] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.346055] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141438, 'name': Rename_Task, 'duration_secs': 0.166029} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.346709] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 945.347322] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8d39709b-e9b6-40d5-9344-2ca7003fd9c2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.359402] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 945.359402] env[62692]: value = "task-1141440" [ 945.359402] env[62692]: _type = "Task" [ 945.359402] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.368967] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141440, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.462836] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3f5dd9de-6b91-4a69-a494-029c28bbe2e4 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Lock "e2b3b062-10e2-4a9c-b6b3-f0786de07b79" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.451s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.572059] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b969a10f-f76c-470f-a4dd-e7ebec6a1ae2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.583341] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58a68e79-ad81-4f9e-b1aa-e23ed30bb3f0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.612571] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "refresh_cache-a668a569-a90e-4e59-8d2e-b0225745b500" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.612721] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquired lock "refresh_cache-a668a569-a90e-4e59-8d2e-b0225745b500" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.612877] env[62692]: DEBUG nova.network.neutron [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 945.615062] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-111a645a-e1c1-4fd9-bd51-a86155684ff8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.626183] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83a86818-be8e-45dd-91b8-db6892be0479 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.644153] env[62692]: DEBUG nova.compute.provider_tree [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.727590] env[62692]: DEBUG oslo_vmware.api [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141439, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074135} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.727884] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 945.728704] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-846f44d2-7aca-4c86-ac30-c8b60a12da2b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.754639] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 3f6c9744-a6e3-43f3-8b6b-624ba178747d/3f6c9744-a6e3-43f3-8b6b-624ba178747d.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 945.755161] env[62692]: INFO nova.compute.manager [-] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Took 1.55 seconds to deallocate network for instance. [ 945.755411] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-51bed0eb-e5b5-44b1-b215-dfb3cf9c408e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.778048] env[62692]: DEBUG oslo_vmware.api [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 945.778048] env[62692]: value = "task-1141441" [ 945.778048] env[62692]: _type = "Task" [ 945.778048] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.787312] env[62692]: DEBUG oslo_vmware.api [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141441, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.872514] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141440, 'name': PowerOnVM_Task} progress is 76%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.888478] env[62692]: DEBUG nova.compute.manager [req-92bb89f2-abd4-45df-8370-8378607e7c01 req-72094e24-1da1-4b14-a957-5405a2250648 service nova] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Received event network-vif-deleted-354f1e83-aa2d-441f-8461-937a689e16c2 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 945.888712] env[62692]: DEBUG nova.compute.manager [req-92bb89f2-abd4-45df-8370-8378607e7c01 req-72094e24-1da1-4b14-a957-5405a2250648 service nova] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Received event network-vif-plugged-f1187f72-32ea-484c-9bde-f40c9a5a3a03 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 945.889052] env[62692]: DEBUG oslo_concurrency.lockutils [req-92bb89f2-abd4-45df-8370-8378607e7c01 req-72094e24-1da1-4b14-a957-5405a2250648 service nova] Acquiring lock "a668a569-a90e-4e59-8d2e-b0225745b500-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.889289] env[62692]: DEBUG oslo_concurrency.lockutils [req-92bb89f2-abd4-45df-8370-8378607e7c01 req-72094e24-1da1-4b14-a957-5405a2250648 service nova] Lock "a668a569-a90e-4e59-8d2e-b0225745b500-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.889514] env[62692]: DEBUG oslo_concurrency.lockutils [req-92bb89f2-abd4-45df-8370-8378607e7c01 req-72094e24-1da1-4b14-a957-5405a2250648 service nova] Lock "a668a569-a90e-4e59-8d2e-b0225745b500-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.889717] env[62692]: DEBUG nova.compute.manager [req-92bb89f2-abd4-45df-8370-8378607e7c01 req-72094e24-1da1-4b14-a957-5405a2250648 service nova] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] No waiting events found dispatching network-vif-plugged-f1187f72-32ea-484c-9bde-f40c9a5a3a03 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 945.889902] env[62692]: WARNING nova.compute.manager [req-92bb89f2-abd4-45df-8370-8378607e7c01 req-72094e24-1da1-4b14-a957-5405a2250648 service nova] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Received unexpected event network-vif-plugged-f1187f72-32ea-484c-9bde-f40c9a5a3a03 for instance with vm_state building and task_state spawning. [ 945.890082] env[62692]: DEBUG nova.compute.manager [req-92bb89f2-abd4-45df-8370-8378607e7c01 req-72094e24-1da1-4b14-a957-5405a2250648 service nova] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Received event network-changed-f1187f72-32ea-484c-9bde-f40c9a5a3a03 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 945.890286] env[62692]: DEBUG nova.compute.manager [req-92bb89f2-abd4-45df-8370-8378607e7c01 req-72094e24-1da1-4b14-a957-5405a2250648 service nova] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Refreshing instance network info cache due to event network-changed-f1187f72-32ea-484c-9bde-f40c9a5a3a03. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 945.890457] env[62692]: DEBUG oslo_concurrency.lockutils [req-92bb89f2-abd4-45df-8370-8378607e7c01 req-72094e24-1da1-4b14-a957-5405a2250648 service nova] Acquiring lock "refresh_cache-a668a569-a90e-4e59-8d2e-b0225745b500" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.149457] env[62692]: DEBUG nova.scheduler.client.report [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 946.162363] env[62692]: DEBUG nova.network.neutron [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 946.276266] env[62692]: DEBUG oslo_concurrency.lockutils [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.291357] env[62692]: DEBUG oslo_vmware.api [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141441, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.330167] env[62692]: DEBUG nova.network.neutron [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Updating instance_info_cache with network_info: [{"id": "f1187f72-32ea-484c-9bde-f40c9a5a3a03", "address": "fa:16:3e:a6:f1:8d", "network": {"id": "2febb660-329d-403c-8c2c-f11c0f5f547c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1806650357-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37921f67664544478cb2568def068b11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1187f72-32", "ovs_interfaceid": "f1187f72-32ea-484c-9bde-f40c9a5a3a03", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.372212] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141440, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.656670] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.529s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.656670] env[62692]: DEBUG nova.compute.manager [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 946.665493] env[62692]: DEBUG oslo_concurrency.lockutils [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.901s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.669286] env[62692]: INFO nova.compute.claims [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 946.791845] env[62692]: DEBUG oslo_vmware.api [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141441, 'name': ReconfigVM_Task, 'duration_secs': 0.914886} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.791845] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 3f6c9744-a6e3-43f3-8b6b-624ba178747d/3f6c9744-a6e3-43f3-8b6b-624ba178747d.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 946.791845] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-df2c127f-0ac4-4840-ada1-60d5f5d1e50c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.800862] env[62692]: DEBUG oslo_vmware.api [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 946.800862] env[62692]: value = "task-1141442" [ 946.800862] env[62692]: _type = "Task" [ 946.800862] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.812965] env[62692]: DEBUG oslo_vmware.api [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141442, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.835064] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Releasing lock "refresh_cache-a668a569-a90e-4e59-8d2e-b0225745b500" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.835388] env[62692]: DEBUG nova.compute.manager [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Instance network_info: |[{"id": "f1187f72-32ea-484c-9bde-f40c9a5a3a03", "address": "fa:16:3e:a6:f1:8d", "network": {"id": "2febb660-329d-403c-8c2c-f11c0f5f547c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1806650357-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37921f67664544478cb2568def068b11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1187f72-32", "ovs_interfaceid": "f1187f72-32ea-484c-9bde-f40c9a5a3a03", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 946.835713] env[62692]: DEBUG oslo_concurrency.lockutils [req-92bb89f2-abd4-45df-8370-8378607e7c01 req-72094e24-1da1-4b14-a957-5405a2250648 service nova] Acquired lock "refresh_cache-a668a569-a90e-4e59-8d2e-b0225745b500" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.835903] env[62692]: DEBUG nova.network.neutron [req-92bb89f2-abd4-45df-8370-8378607e7c01 req-72094e24-1da1-4b14-a957-5405a2250648 service nova] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Refreshing network info cache for port f1187f72-32ea-484c-9bde-f40c9a5a3a03 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 946.837343] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a6:f1:8d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f1187f72-32ea-484c-9bde-f40c9a5a3a03', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 946.849029] env[62692]: DEBUG oslo.service.loopingcall [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 946.852351] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 946.852927] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-77485c15-40fc-49ad-adea-0f1013728fcf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.877847] env[62692]: DEBUG oslo_vmware.api [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141440, 'name': PowerOnVM_Task, 'duration_secs': 1.116716} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.880907] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 946.881010] env[62692]: DEBUG nova.compute.manager [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 946.881383] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 946.881383] env[62692]: value = "task-1141443" [ 946.881383] env[62692]: _type = "Task" [ 946.881383] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.882142] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14a12efd-103a-4bc3-86f3-1d4ab150dcc7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.906659] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141443, 'name': CreateVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.167400] env[62692]: DEBUG nova.compute.utils [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 947.172076] env[62692]: DEBUG nova.compute.manager [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 947.172076] env[62692]: DEBUG nova.network.neutron [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 947.291805] env[62692]: DEBUG nova.policy [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6e9cd6d33e744da4a2f678226b684845', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1037e5d436bd429391159b87a712f6e1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 947.316731] env[62692]: DEBUG oslo_vmware.api [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141442, 'name': Rename_Task, 'duration_secs': 0.188832} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.318225] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 947.318635] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-26f03604-67b4-436d-aeb6-e0231fe90f79 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.327897] env[62692]: DEBUG oslo_vmware.api [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 947.327897] env[62692]: value = "task-1141444" [ 947.327897] env[62692]: _type = "Task" [ 947.327897] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.343896] env[62692]: DEBUG oslo_vmware.api [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141444, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.375318] env[62692]: DEBUG nova.network.neutron [req-92bb89f2-abd4-45df-8370-8378607e7c01 req-72094e24-1da1-4b14-a957-5405a2250648 service nova] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Updated VIF entry in instance network info cache for port f1187f72-32ea-484c-9bde-f40c9a5a3a03. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 947.375465] env[62692]: DEBUG nova.network.neutron [req-92bb89f2-abd4-45df-8370-8378607e7c01 req-72094e24-1da1-4b14-a957-5405a2250648 service nova] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Updating instance_info_cache with network_info: [{"id": "f1187f72-32ea-484c-9bde-f40c9a5a3a03", "address": "fa:16:3e:a6:f1:8d", "network": {"id": "2febb660-329d-403c-8c2c-f11c0f5f547c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1806650357-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37921f67664544478cb2568def068b11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1187f72-32", "ovs_interfaceid": "f1187f72-32ea-484c-9bde-f40c9a5a3a03", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.402898] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141443, 'name': CreateVM_Task, 'duration_secs': 0.461996} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.403069] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 947.403833] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.404131] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.404374] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 947.404687] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a1f1d04-97b6-4d3d-b92e-467a2776a921 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.415400] env[62692]: DEBUG oslo_vmware.api [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 947.415400] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52acbf84-f208-da5a-44f2-e4c2001113e0" [ 947.415400] env[62692]: _type = "Task" [ 947.415400] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.417022] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.425460] env[62692]: DEBUG oslo_vmware.api [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52acbf84-f208-da5a-44f2-e4c2001113e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.578733] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Acquiring lock "e2b3b062-10e2-4a9c-b6b3-f0786de07b79" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.579040] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Lock "e2b3b062-10e2-4a9c-b6b3-f0786de07b79" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.579318] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Acquiring lock "e2b3b062-10e2-4a9c-b6b3-f0786de07b79-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.579682] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Lock "e2b3b062-10e2-4a9c-b6b3-f0786de07b79-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.579682] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Lock "e2b3b062-10e2-4a9c-b6b3-f0786de07b79-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.582364] env[62692]: INFO nova.compute.manager [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Terminating instance [ 947.585971] env[62692]: DEBUG nova.compute.manager [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 947.586230] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 947.587090] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90f1fe34-15b7-4554-9d3c-82b9113d3c19 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.596189] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 947.596867] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-499e1848-a01d-46e9-adea-c71861c62e3f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.603915] env[62692]: DEBUG oslo_vmware.api [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Waiting for the task: (returnval){ [ 947.603915] env[62692]: value = "task-1141445" [ 947.603915] env[62692]: _type = "Task" [ 947.603915] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.614806] env[62692]: DEBUG oslo_vmware.api [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Task: {'id': task-1141445, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.673381] env[62692]: DEBUG nova.compute.manager [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 947.743637] env[62692]: DEBUG nova.network.neutron [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Successfully created port: 5ff6939b-f0fd-4b3d-b0e8-f7946f80493b {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 947.846483] env[62692]: DEBUG oslo_vmware.api [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141444, 'name': PowerOnVM_Task} progress is 90%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.883235] env[62692]: DEBUG oslo_concurrency.lockutils [req-92bb89f2-abd4-45df-8370-8378607e7c01 req-72094e24-1da1-4b14-a957-5405a2250648 service nova] Releasing lock "refresh_cache-a668a569-a90e-4e59-8d2e-b0225745b500" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.926745] env[62692]: DEBUG oslo_vmware.api [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52acbf84-f208-da5a-44f2-e4c2001113e0, 'name': SearchDatastore_Task, 'duration_secs': 0.035203} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.930160] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.930160] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 947.930262] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.930401] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.930581] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 947.931199] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e3567c0c-2b8e-489a-b0f1-fa1c188d5d9f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.947855] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 947.948223] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 947.949104] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3bc7a00b-c74a-49e8-aa73-3d78546d5f6b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.957022] env[62692]: DEBUG oslo_vmware.api [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 947.957022] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]523d248d-cba0-f712-7deb-0f45e3d05b5b" [ 947.957022] env[62692]: _type = "Task" [ 947.957022] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.969538] env[62692]: DEBUG oslo_vmware.api [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]523d248d-cba0-f712-7deb-0f45e3d05b5b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.118979] env[62692]: DEBUG oslo_vmware.api [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Task: {'id': task-1141445, 'name': PowerOffVM_Task, 'duration_secs': 0.249902} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.119697] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 948.119697] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 948.119922] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c050a69c-b940-4b85-bc2d-b9e7fe455b4f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.188898] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d79d4068-c99e-4887-bc71-bf6d2a70ad32 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.204970] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 948.205286] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 948.205402] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Deleting the datastore file [datastore1] e2b3b062-10e2-4a9c-b6b3-f0786de07b79 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 948.208453] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d017c470-b712-42a4-9d3e-0aaca2218dcc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.212695] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1f2c3028-db92-4e76-8030-e92eac4cc8fe {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.245476] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cf44bc5-0362-417a-9046-4ba3a47b28f8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.248350] env[62692]: DEBUG oslo_vmware.api [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Waiting for the task: (returnval){ [ 948.248350] env[62692]: value = "task-1141447" [ 948.248350] env[62692]: _type = "Task" [ 948.248350] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.255428] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5979c0ae-c833-49db-bc57-f6187b4af7e3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.262358] env[62692]: DEBUG oslo_vmware.api [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Task: {'id': task-1141447, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.272894] env[62692]: DEBUG nova.compute.provider_tree [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.291874] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5e56b379-8fa6-4d66-b63d-c83f10ce4265 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.292133] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5e56b379-8fa6-4d66-b63d-c83f10ce4265 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.342794] env[62692]: DEBUG oslo_vmware.api [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141444, 'name': PowerOnVM_Task, 'duration_secs': 0.634943} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.343698] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 948.343698] env[62692]: INFO nova.compute.manager [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Took 8.94 seconds to spawn the instance on the hypervisor. [ 948.343698] env[62692]: DEBUG nova.compute.manager [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 948.344704] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b332c335-016d-4c6e-a293-a97e87bcbb3d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.467752] env[62692]: DEBUG oslo_vmware.api [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]523d248d-cba0-f712-7deb-0f45e3d05b5b, 'name': SearchDatastore_Task, 'duration_secs': 0.032105} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.468588] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bde0aa78-a0fa-42e5-bd32-cec8f86359f4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.474356] env[62692]: DEBUG oslo_vmware.api [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 948.474356] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52771af3-732f-7233-f666-2ca05e6cf39d" [ 948.474356] env[62692]: _type = "Task" [ 948.474356] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.482980] env[62692]: DEBUG oslo_vmware.api [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52771af3-732f-7233-f666-2ca05e6cf39d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.685158] env[62692]: DEBUG nova.compute.manager [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 948.715358] env[62692]: DEBUG nova.virt.hardware [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 948.715358] env[62692]: DEBUG nova.virt.hardware [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 948.715502] env[62692]: DEBUG nova.virt.hardware [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 948.715681] env[62692]: DEBUG nova.virt.hardware [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 948.715843] env[62692]: DEBUG nova.virt.hardware [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 948.716149] env[62692]: DEBUG nova.virt.hardware [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 948.716350] env[62692]: DEBUG nova.virt.hardware [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 948.716597] env[62692]: DEBUG nova.virt.hardware [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 948.716850] env[62692]: DEBUG nova.virt.hardware [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 948.717303] env[62692]: DEBUG nova.virt.hardware [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 948.717484] env[62692]: DEBUG nova.virt.hardware [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 948.718422] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fd1456e-2a46-4557-8178-8c9080716b9e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.727163] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77aa26ec-9129-455d-a2d1-463bbc8dda9a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.757340] env[62692]: DEBUG oslo_vmware.api [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Task: {'id': task-1141447, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.455592} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.757604] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 948.757834] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 948.758037] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 948.758224] env[62692]: INFO nova.compute.manager [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Took 1.17 seconds to destroy the instance on the hypervisor. [ 948.758468] env[62692]: DEBUG oslo.service.loopingcall [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 948.758659] env[62692]: DEBUG nova.compute.manager [-] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 948.758754] env[62692]: DEBUG nova.network.neutron [-] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 948.775453] env[62692]: DEBUG nova.scheduler.client.report [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 948.794953] env[62692]: DEBUG nova.compute.utils [None req-5e56b379-8fa6-4d66-b63d-c83f10ce4265 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 948.863609] env[62692]: INFO nova.compute.manager [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Rebuilding instance [ 948.868355] env[62692]: INFO nova.compute.manager [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Took 37.47 seconds to build instance. [ 948.906599] env[62692]: DEBUG nova.compute.manager [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 948.907502] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7af0323-07bf-4101-8499-ff17ef304731 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.986762] env[62692]: DEBUG oslo_vmware.api [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52771af3-732f-7233-f666-2ca05e6cf39d, 'name': SearchDatastore_Task, 'duration_secs': 0.017341} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.987050] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.987322] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] a668a569-a90e-4e59-8d2e-b0225745b500/a668a569-a90e-4e59-8d2e-b0225745b500.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 948.987584] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-83d9d7a2-08b5-4658-84b6-57a34ee57612 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.994651] env[62692]: DEBUG oslo_vmware.api [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 948.994651] env[62692]: value = "task-1141448" [ 948.994651] env[62692]: _type = "Task" [ 948.994651] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.007826] env[62692]: DEBUG oslo_vmware.api [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141448, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.283018] env[62692]: DEBUG oslo_concurrency.lockutils [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.615s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.283018] env[62692]: DEBUG nova.compute.manager [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 949.285180] env[62692]: DEBUG oslo_concurrency.lockutils [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.856s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.294031] env[62692]: INFO nova.compute.claims [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 949.299074] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5e56b379-8fa6-4d66-b63d-c83f10ce4265 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.371960] env[62692]: DEBUG oslo_concurrency.lockutils [None req-595bfb9f-3a45-487b-b535-106bb7b172ae tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "3f6c9744-a6e3-43f3-8b6b-624ba178747d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.111s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.423024] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 949.423024] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3d8a2eb4-97a0-46e2-8b40-bd6834032561 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.431337] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 949.431337] env[62692]: value = "task-1141449" [ 949.431337] env[62692]: _type = "Task" [ 949.431337] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.442881] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141449, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.509661] env[62692]: DEBUG oslo_vmware.api [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141448, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.518320] env[62692]: DEBUG nova.network.neutron [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Successfully updated port: 5ff6939b-f0fd-4b3d-b0e8-f7946f80493b {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 949.530027] env[62692]: DEBUG nova.network.neutron [-] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.540518] env[62692]: DEBUG nova.compute.manager [req-3de5532d-5219-40c4-93a0-c7cab95ddd5b req-f9aba08b-058b-4c7a-87df-cf99653637ef service nova] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Received event network-vif-deleted-62803691-31f7-422f-b373-a85721fe22b6 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 949.540724] env[62692]: INFO nova.compute.manager [req-3de5532d-5219-40c4-93a0-c7cab95ddd5b req-f9aba08b-058b-4c7a-87df-cf99653637ef service nova] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Neutron deleted interface 62803691-31f7-422f-b373-a85721fe22b6; detaching it from the instance and deleting it from the info cache [ 949.540898] env[62692]: DEBUG nova.network.neutron [req-3de5532d-5219-40c4-93a0-c7cab95ddd5b req-f9aba08b-058b-4c7a-87df-cf99653637ef service nova] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.800568] env[62692]: DEBUG nova.compute.utils [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 949.803824] env[62692]: DEBUG nova.compute.manager [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 949.804009] env[62692]: DEBUG nova.network.neutron [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 949.882033] env[62692]: DEBUG nova.policy [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '095cfec4d2ed41498767836c72b13915', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4b111e98b02a401384b23e363372ed52', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 949.943565] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141449, 'name': PowerOffVM_Task, 'duration_secs': 0.207164} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.943896] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 949.945050] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 949.945835] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adf569af-4075-48fd-b4ef-a095577a92d9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.953647] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 949.953882] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f450e9ce-6960-40e6-a05d-f3b7d6d57748 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.005676] env[62692]: DEBUG oslo_vmware.api [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141448, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.581381} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.005955] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] a668a569-a90e-4e59-8d2e-b0225745b500/a668a569-a90e-4e59-8d2e-b0225745b500.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 950.006200] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 950.006484] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-53b0d6a8-0fb7-43d4-b6aa-31a2d99113a0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.015171] env[62692]: DEBUG oslo_vmware.api [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 950.015171] env[62692]: value = "task-1141451" [ 950.015171] env[62692]: _type = "Task" [ 950.015171] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.021468] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "refresh_cache-46bce865-188b-487c-a73f-cf524059c797" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.021616] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired lock "refresh_cache-46bce865-188b-487c-a73f-cf524059c797" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.021784] env[62692]: DEBUG nova.network.neutron [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 950.026834] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 950.027860] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 950.027860] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Deleting the datastore file [datastore1] 5a08fa8a-f9fe-4879-bb7b-baa04097df6c {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 950.027860] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e6cb74cc-cc22-46d8-9daf-8271387cf4d9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.036965] env[62692]: DEBUG oslo_vmware.api [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141451, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.037894] env[62692]: INFO nova.compute.manager [-] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Took 1.28 seconds to deallocate network for instance. [ 950.045439] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1392ea8d-1e96-4976-bf1e-cfe0decaeb71 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.047486] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 950.047486] env[62692]: value = "task-1141452" [ 950.047486] env[62692]: _type = "Task" [ 950.047486] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.056023] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-135d6976-931a-4551-9be7-66ce35552eaf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.075717] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141452, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.096930] env[62692]: DEBUG nova.compute.manager [req-3de5532d-5219-40c4-93a0-c7cab95ddd5b req-f9aba08b-058b-4c7a-87df-cf99653637ef service nova] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Detach interface failed, port_id=62803691-31f7-422f-b373-a85721fe22b6, reason: Instance e2b3b062-10e2-4a9c-b6b3-f0786de07b79 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 950.242438] env[62692]: DEBUG nova.network.neutron [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Successfully created port: 3fa9bd6e-092d-4e9f-a4ab-5113affb920b {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 950.310766] env[62692]: DEBUG nova.compute.manager [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 950.369097] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5e56b379-8fa6-4d66-b63d-c83f10ce4265 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.369097] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5e56b379-8fa6-4d66-b63d-c83f10ce4265 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.369259] env[62692]: INFO nova.compute.manager [None req-5e56b379-8fa6-4d66-b63d-c83f10ce4265 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Attaching volume 50c399be-767e-458d-bf5a-9c2395ab8e24 to /dev/sdb [ 950.413195] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a089805-c408-410b-8548-6746c873ae8e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.421461] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83e44408-d15f-42d6-9047-611ce8de5800 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.436064] env[62692]: DEBUG nova.virt.block_device [None req-5e56b379-8fa6-4d66-b63d-c83f10ce4265 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Updating existing volume attachment record: ede9a949-4b26-4075-98bf-349e4f383773 {{(pid=62692) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 950.524984] env[62692]: DEBUG oslo_vmware.api [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141451, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080868} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.527553] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 950.531545] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75eef800-20c4-4de1-ab36-2cac21249427 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.546476] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.556040] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] a668a569-a90e-4e59-8d2e-b0225745b500/a668a569-a90e-4e59-8d2e-b0225745b500.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 950.559019] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1b63840-bbe2-45a7-919b-214960194219 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.583687] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141452, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.500784} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.584537] env[62692]: DEBUG nova.network.neutron [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 950.592026] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 950.592026] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 950.592026] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 950.593296] env[62692]: DEBUG oslo_vmware.api [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 950.593296] env[62692]: value = "task-1141453" [ 950.593296] env[62692]: _type = "Task" [ 950.593296] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.602742] env[62692]: DEBUG oslo_vmware.api [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141453, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.803855] env[62692]: DEBUG nova.network.neutron [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Updating instance_info_cache with network_info: [{"id": "5ff6939b-f0fd-4b3d-b0e8-f7946f80493b", "address": "fa:16:3e:04:13:61", "network": {"id": "f70727d6-a9aa-44b0-a244-2533b3f96571", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-284105837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1037e5d436bd429391159b87a712f6e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ff6939b-f0", "ovs_interfaceid": "5ff6939b-f0fd-4b3d-b0e8-f7946f80493b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.809032] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-218ae73f-8ef3-4aa1-b9a2-c76d8921dae7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.821996] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dc95ec3-1b2a-4c43-a938-3a6d2f25a4cf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.856249] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7d370a2-5ee4-4b01-9a8e-ad7bde7b8cd9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.868413] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df31959c-fe74-485a-807b-3c3464c2ee39 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.886378] env[62692]: DEBUG nova.compute.provider_tree [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 951.114177] env[62692]: DEBUG oslo_vmware.api [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141453, 'name': ReconfigVM_Task, 'duration_secs': 0.331007} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.114528] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Reconfigured VM instance instance-0000004c to attach disk [datastore1] a668a569-a90e-4e59-8d2e-b0225745b500/a668a569-a90e-4e59-8d2e-b0225745b500.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 951.115593] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-286a0fa3-6af0-4e20-a854-3009786bdcd6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.123875] env[62692]: DEBUG oslo_vmware.api [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 951.123875] env[62692]: value = "task-1141457" [ 951.123875] env[62692]: _type = "Task" [ 951.123875] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.134318] env[62692]: DEBUG oslo_vmware.api [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141457, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.307743] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Releasing lock "refresh_cache-46bce865-188b-487c-a73f-cf524059c797" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.308101] env[62692]: DEBUG nova.compute.manager [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Instance network_info: |[{"id": "5ff6939b-f0fd-4b3d-b0e8-f7946f80493b", "address": "fa:16:3e:04:13:61", "network": {"id": "f70727d6-a9aa-44b0-a244-2533b3f96571", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-284105837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1037e5d436bd429391159b87a712f6e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ff6939b-f0", "ovs_interfaceid": "5ff6939b-f0fd-4b3d-b0e8-f7946f80493b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 951.308561] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:04:13:61', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f44b2fa3-6730-4b87-8839-947eff21213f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5ff6939b-f0fd-4b3d-b0e8-f7946f80493b', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 951.316482] env[62692]: DEBUG oslo.service.loopingcall [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 951.316724] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 46bce865-188b-487c-a73f-cf524059c797] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 951.316959] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e2fd3340-b9c6-494c-ae48-42739e26f814 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.333028] env[62692]: DEBUG nova.compute.manager [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 951.342584] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 951.342584] env[62692]: value = "task-1141458" [ 951.342584] env[62692]: _type = "Task" [ 951.342584] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.356026] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141458, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.371547] env[62692]: DEBUG nova.virt.hardware [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 951.372878] env[62692]: DEBUG nova.virt.hardware [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 951.372878] env[62692]: DEBUG nova.virt.hardware [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 951.372878] env[62692]: DEBUG nova.virt.hardware [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 951.372878] env[62692]: DEBUG nova.virt.hardware [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 951.372878] env[62692]: DEBUG nova.virt.hardware [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 951.372878] env[62692]: DEBUG nova.virt.hardware [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 951.372878] env[62692]: DEBUG nova.virt.hardware [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 951.373187] env[62692]: DEBUG nova.virt.hardware [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 951.373240] env[62692]: DEBUG nova.virt.hardware [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 951.373460] env[62692]: DEBUG nova.virt.hardware [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 951.374364] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f749152f-8b6a-4a70-9001-aa02fe49810d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.382279] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bb99634-b16f-43cb-87a0-0cecdc2e06f2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.396569] env[62692]: DEBUG nova.scheduler.client.report [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 951.569839] env[62692]: DEBUG nova.compute.manager [req-77f662ee-a0e9-49d5-ba88-2f9031bacdf6 req-93cf42dd-5637-4762-8368-8c318b61ec3e service nova] [instance: 46bce865-188b-487c-a73f-cf524059c797] Received event network-vif-plugged-5ff6939b-f0fd-4b3d-b0e8-f7946f80493b {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 951.570234] env[62692]: DEBUG oslo_concurrency.lockutils [req-77f662ee-a0e9-49d5-ba88-2f9031bacdf6 req-93cf42dd-5637-4762-8368-8c318b61ec3e service nova] Acquiring lock "46bce865-188b-487c-a73f-cf524059c797-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.570522] env[62692]: DEBUG oslo_concurrency.lockutils [req-77f662ee-a0e9-49d5-ba88-2f9031bacdf6 req-93cf42dd-5637-4762-8368-8c318b61ec3e service nova] Lock "46bce865-188b-487c-a73f-cf524059c797-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.570707] env[62692]: DEBUG oslo_concurrency.lockutils [req-77f662ee-a0e9-49d5-ba88-2f9031bacdf6 req-93cf42dd-5637-4762-8368-8c318b61ec3e service nova] Lock "46bce865-188b-487c-a73f-cf524059c797-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.570884] env[62692]: DEBUG nova.compute.manager [req-77f662ee-a0e9-49d5-ba88-2f9031bacdf6 req-93cf42dd-5637-4762-8368-8c318b61ec3e service nova] [instance: 46bce865-188b-487c-a73f-cf524059c797] No waiting events found dispatching network-vif-plugged-5ff6939b-f0fd-4b3d-b0e8-f7946f80493b {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 951.571069] env[62692]: WARNING nova.compute.manager [req-77f662ee-a0e9-49d5-ba88-2f9031bacdf6 req-93cf42dd-5637-4762-8368-8c318b61ec3e service nova] [instance: 46bce865-188b-487c-a73f-cf524059c797] Received unexpected event network-vif-plugged-5ff6939b-f0fd-4b3d-b0e8-f7946f80493b for instance with vm_state building and task_state spawning. [ 951.571263] env[62692]: DEBUG nova.compute.manager [req-77f662ee-a0e9-49d5-ba88-2f9031bacdf6 req-93cf42dd-5637-4762-8368-8c318b61ec3e service nova] [instance: 46bce865-188b-487c-a73f-cf524059c797] Received event network-changed-5ff6939b-f0fd-4b3d-b0e8-f7946f80493b {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 951.571509] env[62692]: DEBUG nova.compute.manager [req-77f662ee-a0e9-49d5-ba88-2f9031bacdf6 req-93cf42dd-5637-4762-8368-8c318b61ec3e service nova] [instance: 46bce865-188b-487c-a73f-cf524059c797] Refreshing instance network info cache due to event network-changed-5ff6939b-f0fd-4b3d-b0e8-f7946f80493b. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 951.571768] env[62692]: DEBUG oslo_concurrency.lockutils [req-77f662ee-a0e9-49d5-ba88-2f9031bacdf6 req-93cf42dd-5637-4762-8368-8c318b61ec3e service nova] Acquiring lock "refresh_cache-46bce865-188b-487c-a73f-cf524059c797" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.571925] env[62692]: DEBUG oslo_concurrency.lockutils [req-77f662ee-a0e9-49d5-ba88-2f9031bacdf6 req-93cf42dd-5637-4762-8368-8c318b61ec3e service nova] Acquired lock "refresh_cache-46bce865-188b-487c-a73f-cf524059c797" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.572104] env[62692]: DEBUG nova.network.neutron [req-77f662ee-a0e9-49d5-ba88-2f9031bacdf6 req-93cf42dd-5637-4762-8368-8c318b61ec3e service nova] [instance: 46bce865-188b-487c-a73f-cf524059c797] Refreshing network info cache for port 5ff6939b-f0fd-4b3d-b0e8-f7946f80493b {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 951.631586] env[62692]: DEBUG nova.virt.hardware [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 951.631929] env[62692]: DEBUG nova.virt.hardware [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 951.632162] env[62692]: DEBUG nova.virt.hardware [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 951.633237] env[62692]: DEBUG nova.virt.hardware [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 951.633421] env[62692]: DEBUG nova.virt.hardware [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 951.633584] env[62692]: DEBUG nova.virt.hardware [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 951.633805] env[62692]: DEBUG nova.virt.hardware [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 951.633976] env[62692]: DEBUG nova.virt.hardware [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 951.634172] env[62692]: DEBUG nova.virt.hardware [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 951.634346] env[62692]: DEBUG nova.virt.hardware [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 951.634530] env[62692]: DEBUG nova.virt.hardware [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 951.635407] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89f3809c-ca5c-413a-acca-3e3195d10a1c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.641220] env[62692]: DEBUG oslo_vmware.api [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141457, 'name': Rename_Task, 'duration_secs': 0.165489} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.641794] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 951.642048] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-63002664-87e9-4188-be43-f82523d0f142 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.648221] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b30737be-85bf-492c-ba4b-30a34e4886e1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.653987] env[62692]: DEBUG oslo_vmware.api [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 951.653987] env[62692]: value = "task-1141459" [ 951.653987] env[62692]: _type = "Task" [ 951.653987] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.665909] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6c:33:e3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '678ebbe4-4c53-4eaf-a689-93981310f37d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6f540034-8dda-4def-adb2-0cdba3268510', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 951.673481] env[62692]: DEBUG oslo.service.loopingcall [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 951.674282] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 951.674534] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1eb1cf98-0b5c-40aa-b0b9-42e1ab4939f9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.692688] env[62692]: DEBUG oslo_vmware.api [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141459, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.699175] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 951.699175] env[62692]: value = "task-1141460" [ 951.699175] env[62692]: _type = "Task" [ 951.699175] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.716856] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141460, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.854582] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141458, 'name': CreateVM_Task, 'duration_secs': 0.462658} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.854685] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 46bce865-188b-487c-a73f-cf524059c797] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 951.855497] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.855669] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.855990] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 951.856376] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c53c6f23-cd7c-46d1-9f0b-053a698a4b7a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.862081] env[62692]: DEBUG oslo_vmware.api [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 951.862081] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52813a7b-1a38-4a9d-b12e-758eefd21a64" [ 951.862081] env[62692]: _type = "Task" [ 951.862081] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.870914] env[62692]: DEBUG oslo_vmware.api [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52813a7b-1a38-4a9d-b12e-758eefd21a64, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.902098] env[62692]: DEBUG oslo_concurrency.lockutils [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.617s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.902680] env[62692]: DEBUG nova.compute.manager [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 951.905476] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.142s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.905696] env[62692]: DEBUG nova.objects.instance [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Lazy-loading 'resources' on Instance uuid 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 951.969810] env[62692]: DEBUG nova.network.neutron [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Successfully updated port: 3fa9bd6e-092d-4e9f-a4ab-5113affb920b {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 952.166352] env[62692]: DEBUG oslo_vmware.api [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141459, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.210811] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141460, 'name': CreateVM_Task, 'duration_secs': 0.459846} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.213657] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 952.214058] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.214244] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.214570] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 952.215146] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc970e32-f503-49f5-baff-f6d39ff5a16e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.221567] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 952.221567] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52411f1d-00ab-8b48-6996-8d0e2bc66e13" [ 952.221567] env[62692]: _type = "Task" [ 952.221567] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.224651] env[62692]: DEBUG nova.compute.manager [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Stashing vm_state: active {{(pid=62692) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 952.237890] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52411f1d-00ab-8b48-6996-8d0e2bc66e13, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.359810] env[62692]: DEBUG nova.network.neutron [req-77f662ee-a0e9-49d5-ba88-2f9031bacdf6 req-93cf42dd-5637-4762-8368-8c318b61ec3e service nova] [instance: 46bce865-188b-487c-a73f-cf524059c797] Updated VIF entry in instance network info cache for port 5ff6939b-f0fd-4b3d-b0e8-f7946f80493b. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 952.360265] env[62692]: DEBUG nova.network.neutron [req-77f662ee-a0e9-49d5-ba88-2f9031bacdf6 req-93cf42dd-5637-4762-8368-8c318b61ec3e service nova] [instance: 46bce865-188b-487c-a73f-cf524059c797] Updating instance_info_cache with network_info: [{"id": "5ff6939b-f0fd-4b3d-b0e8-f7946f80493b", "address": "fa:16:3e:04:13:61", "network": {"id": "f70727d6-a9aa-44b0-a244-2533b3f96571", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-284105837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1037e5d436bd429391159b87a712f6e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ff6939b-f0", "ovs_interfaceid": "5ff6939b-f0fd-4b3d-b0e8-f7946f80493b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.375073] env[62692]: DEBUG oslo_vmware.api [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52813a7b-1a38-4a9d-b12e-758eefd21a64, 'name': SearchDatastore_Task, 'duration_secs': 0.013498} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.375410] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.376016] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 952.376282] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.376441] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.376630] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 952.377020] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c8a748bf-aa67-40ad-9359-cafb41b8d2a9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.392950] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 952.393174] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 952.393990] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-443210e3-f121-490c-a4f0-a0970448c0a0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.399927] env[62692]: DEBUG oslo_vmware.api [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 952.399927] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52a25097-01d4-6e99-e69b-4a9f22a83d17" [ 952.399927] env[62692]: _type = "Task" [ 952.399927] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.409166] env[62692]: DEBUG nova.compute.utils [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 952.412684] env[62692]: DEBUG oslo_vmware.api [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52a25097-01d4-6e99-e69b-4a9f22a83d17, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.413357] env[62692]: DEBUG nova.compute.manager [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 952.413591] env[62692]: DEBUG nova.network.neutron [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 952.464633] env[62692]: DEBUG nova.policy [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a251ef5f6cb143178928babc75a0fef0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ba97a775e99a463c845518b0df8cd3e2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 952.472310] env[62692]: DEBUG oslo_concurrency.lockutils [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Acquiring lock "refresh_cache-084ce29c-302a-45cc-b36f-0143eccee298" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.472547] env[62692]: DEBUG oslo_concurrency.lockutils [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Acquired lock "refresh_cache-084ce29c-302a-45cc-b36f-0143eccee298" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.472612] env[62692]: DEBUG nova.network.neutron [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 952.668525] env[62692]: DEBUG oslo_vmware.api [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141459, 'name': PowerOnVM_Task, 'duration_secs': 0.555531} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.669303] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 952.669662] env[62692]: INFO nova.compute.manager [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Took 8.58 seconds to spawn the instance on the hypervisor. [ 952.671108] env[62692]: DEBUG nova.compute.manager [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 952.671108] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b39da5a-9a45-4bbf-ab97-747426cb8862 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.747484] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52411f1d-00ab-8b48-6996-8d0e2bc66e13, 'name': SearchDatastore_Task, 'duration_secs': 0.022639} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.748653] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.749037] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.749280] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 952.749552] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.749866] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.750079] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 952.750639] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-34c48b75-10b2-4c1c-8641-53db10d0542d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.765557] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 952.765557] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 952.765557] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d8d4435-cfa6-43d2-9c2a-c28bc1c5fad3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.779179] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 952.779179] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52e45adc-8940-bd70-f6d5-691934e2cf4d" [ 952.779179] env[62692]: _type = "Task" [ 952.779179] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.792159] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52e45adc-8940-bd70-f6d5-691934e2cf4d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.815030] env[62692]: DEBUG nova.network.neutron [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Successfully created port: 2deb4259-2f64-4f7c-9289-eba76375335a {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 952.863214] env[62692]: DEBUG oslo_concurrency.lockutils [req-77f662ee-a0e9-49d5-ba88-2f9031bacdf6 req-93cf42dd-5637-4762-8368-8c318b61ec3e service nova] Releasing lock "refresh_cache-46bce865-188b-487c-a73f-cf524059c797" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.874124] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6273cdb9-4906-49a7-a3b7-a3aac8db62cd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.882062] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5ba9c94-bec4-4fc1-a770-09293f865399 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.916532] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6dac098-894e-4efe-829e-f9ee30e0fe02 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.919287] env[62692]: DEBUG nova.compute.manager [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 952.934252] env[62692]: DEBUG oslo_vmware.api [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52a25097-01d4-6e99-e69b-4a9f22a83d17, 'name': SearchDatastore_Task, 'duration_secs': 0.028086} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.937769] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec7ac291-37eb-4286-a6a3-9cbc4596a70d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.944616] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-811b3252-37f8-4c42-9354-16e926b141f7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.951964] env[62692]: DEBUG oslo_vmware.api [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 952.951964] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52082f10-9fce-c0f2-cab1-f10d40e54b51" [ 952.951964] env[62692]: _type = "Task" [ 952.951964] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.964256] env[62692]: DEBUG nova.compute.provider_tree [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.971809] env[62692]: DEBUG oslo_vmware.api [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52082f10-9fce-c0f2-cab1-f10d40e54b51, 'name': SearchDatastore_Task, 'duration_secs': 0.012764} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.972756] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.973039] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 46bce865-188b-487c-a73f-cf524059c797/46bce865-188b-487c-a73f-cf524059c797.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 952.973771] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a51afe10-736b-4718-a81f-80e65ca72271 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.986571] env[62692]: DEBUG oslo_vmware.api [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 952.986571] env[62692]: value = "task-1141462" [ 952.986571] env[62692]: _type = "Task" [ 952.986571] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.995069] env[62692]: DEBUG oslo_vmware.api [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141462, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.190812] env[62692]: INFO nova.compute.manager [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Took 35.87 seconds to build instance. [ 953.236648] env[62692]: DEBUG nova.network.neutron [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 953.299546] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52e45adc-8940-bd70-f6d5-691934e2cf4d, 'name': SearchDatastore_Task, 'duration_secs': 0.021123} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.303859] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1a38c25-1a22-47bc-b6b0-6f77d94b0a30 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.312541] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 953.312541] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52ebfa62-8e58-477d-9061-131e51ae6260" [ 953.312541] env[62692]: _type = "Task" [ 953.312541] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.323911] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52ebfa62-8e58-477d-9061-131e51ae6260, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.467950] env[62692]: DEBUG nova.scheduler.client.report [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 953.496594] env[62692]: DEBUG oslo_vmware.api [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141462, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.542151] env[62692]: DEBUG nova.network.neutron [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Updating instance_info_cache with network_info: [{"id": "3fa9bd6e-092d-4e9f-a4ab-5113affb920b", "address": "fa:16:3e:a2:35:15", "network": {"id": "bbdf7bb9-6b54-4e19-91c6-cb0650e1121b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-2011930120-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "4b111e98b02a401384b23e363372ed52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5", "external-id": "nsx-vlan-transportzone-685", "segmentation_id": 685, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3fa9bd6e-09", "ovs_interfaceid": "3fa9bd6e-092d-4e9f-a4ab-5113affb920b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.692604] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ffe5fe2a-5d2b-4156-893b-6b7ae1c96e05 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "a668a569-a90e-4e59-8d2e-b0225745b500" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.965s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.777904] env[62692]: DEBUG nova.compute.manager [req-335437f6-05d7-4d00-a49c-d52d8e53a89e req-b2e9e1e6-6547-4b78-9c8f-75126f2295f7 service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Received event network-vif-plugged-3fa9bd6e-092d-4e9f-a4ab-5113affb920b {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 953.778169] env[62692]: DEBUG oslo_concurrency.lockutils [req-335437f6-05d7-4d00-a49c-d52d8e53a89e req-b2e9e1e6-6547-4b78-9c8f-75126f2295f7 service nova] Acquiring lock "084ce29c-302a-45cc-b36f-0143eccee298-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.778394] env[62692]: DEBUG oslo_concurrency.lockutils [req-335437f6-05d7-4d00-a49c-d52d8e53a89e req-b2e9e1e6-6547-4b78-9c8f-75126f2295f7 service nova] Lock "084ce29c-302a-45cc-b36f-0143eccee298-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.778572] env[62692]: DEBUG oslo_concurrency.lockutils [req-335437f6-05d7-4d00-a49c-d52d8e53a89e req-b2e9e1e6-6547-4b78-9c8f-75126f2295f7 service nova] Lock "084ce29c-302a-45cc-b36f-0143eccee298-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.778783] env[62692]: DEBUG nova.compute.manager [req-335437f6-05d7-4d00-a49c-d52d8e53a89e req-b2e9e1e6-6547-4b78-9c8f-75126f2295f7 service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] No waiting events found dispatching network-vif-plugged-3fa9bd6e-092d-4e9f-a4ab-5113affb920b {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 953.778913] env[62692]: WARNING nova.compute.manager [req-335437f6-05d7-4d00-a49c-d52d8e53a89e req-b2e9e1e6-6547-4b78-9c8f-75126f2295f7 service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Received unexpected event network-vif-plugged-3fa9bd6e-092d-4e9f-a4ab-5113affb920b for instance with vm_state building and task_state spawning. [ 953.779089] env[62692]: DEBUG nova.compute.manager [req-335437f6-05d7-4d00-a49c-d52d8e53a89e req-b2e9e1e6-6547-4b78-9c8f-75126f2295f7 service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Received event network-changed-3fa9bd6e-092d-4e9f-a4ab-5113affb920b {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 953.779272] env[62692]: DEBUG nova.compute.manager [req-335437f6-05d7-4d00-a49c-d52d8e53a89e req-b2e9e1e6-6547-4b78-9c8f-75126f2295f7 service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Refreshing instance network info cache due to event network-changed-3fa9bd6e-092d-4e9f-a4ab-5113affb920b. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 953.779474] env[62692]: DEBUG oslo_concurrency.lockutils [req-335437f6-05d7-4d00-a49c-d52d8e53a89e req-b2e9e1e6-6547-4b78-9c8f-75126f2295f7 service nova] Acquiring lock "refresh_cache-084ce29c-302a-45cc-b36f-0143eccee298" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.827961] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52ebfa62-8e58-477d-9061-131e51ae6260, 'name': SearchDatastore_Task, 'duration_secs': 0.020072} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.828471] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.828736] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 5a08fa8a-f9fe-4879-bb7b-baa04097df6c/5a08fa8a-f9fe-4879-bb7b-baa04097df6c.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 953.828984] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4e0aedd1-490a-47c6-a1f9-28d605c2e027 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.836880] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 953.836880] env[62692]: value = "task-1141463" [ 953.836880] env[62692]: _type = "Task" [ 953.836880] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.846827] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141463, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.941017] env[62692]: DEBUG nova.compute.manager [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 953.967855] env[62692]: DEBUG nova.virt.hardware [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 953.968180] env[62692]: DEBUG nova.virt.hardware [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 953.968378] env[62692]: DEBUG nova.virt.hardware [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 953.968603] env[62692]: DEBUG nova.virt.hardware [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 953.968788] env[62692]: DEBUG nova.virt.hardware [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 953.968969] env[62692]: DEBUG nova.virt.hardware [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 953.969262] env[62692]: DEBUG nova.virt.hardware [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 953.969455] env[62692]: DEBUG nova.virt.hardware [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 953.969635] env[62692]: DEBUG nova.virt.hardware [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 953.969841] env[62692]: DEBUG nova.virt.hardware [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 953.970067] env[62692]: DEBUG nova.virt.hardware [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 953.970977] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7a2f61c-9537-4493-b7e6-db0a2faf64a4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.978096] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.069s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.978096] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.081s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.978096] env[62692]: DEBUG nova.objects.instance [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Lazy-loading 'resources' on Instance uuid f44e2ee2-6321-4237-9644-a19b5e90ae16 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 953.991454] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b581d53-b36e-4cea-8522-ab18bee9a469 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.998797] env[62692]: INFO nova.scheduler.client.report [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Deleted allocations for instance 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e [ 954.013708] env[62692]: DEBUG oslo_vmware.api [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141462, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.642712} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.013708] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 46bce865-188b-487c-a73f-cf524059c797/46bce865-188b-487c-a73f-cf524059c797.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 954.014016] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 954.014311] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-183455f4-e19e-4f25-823a-9b92be59a716 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.022754] env[62692]: DEBUG oslo_vmware.api [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 954.022754] env[62692]: value = "task-1141464" [ 954.022754] env[62692]: _type = "Task" [ 954.022754] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.033283] env[62692]: DEBUG oslo_vmware.api [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141464, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.048587] env[62692]: DEBUG oslo_concurrency.lockutils [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Releasing lock "refresh_cache-084ce29c-302a-45cc-b36f-0143eccee298" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.048961] env[62692]: DEBUG nova.compute.manager [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Instance network_info: |[{"id": "3fa9bd6e-092d-4e9f-a4ab-5113affb920b", "address": "fa:16:3e:a2:35:15", "network": {"id": "bbdf7bb9-6b54-4e19-91c6-cb0650e1121b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-2011930120-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "4b111e98b02a401384b23e363372ed52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5", "external-id": "nsx-vlan-transportzone-685", "segmentation_id": 685, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3fa9bd6e-09", "ovs_interfaceid": "3fa9bd6e-092d-4e9f-a4ab-5113affb920b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 954.049359] env[62692]: DEBUG oslo_concurrency.lockutils [req-335437f6-05d7-4d00-a49c-d52d8e53a89e req-b2e9e1e6-6547-4b78-9c8f-75126f2295f7 service nova] Acquired lock "refresh_cache-084ce29c-302a-45cc-b36f-0143eccee298" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.049574] env[62692]: DEBUG nova.network.neutron [req-335437f6-05d7-4d00-a49c-d52d8e53a89e req-b2e9e1e6-6547-4b78-9c8f-75126f2295f7 service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Refreshing network info cache for port 3fa9bd6e-092d-4e9f-a4ab-5113affb920b {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 954.050958] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a2:35:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3fa9bd6e-092d-4e9f-a4ab-5113affb920b', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 954.058941] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Creating folder: Project (4b111e98b02a401384b23e363372ed52). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 954.060146] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d1c71486-28e5-422a-a445-06b84a6e9cf4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.074037] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Created folder: Project (4b111e98b02a401384b23e363372ed52) in parent group-v248868. [ 954.074303] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Creating folder: Instances. Parent ref: group-v249000. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 954.074864] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b1035783-b6fe-4cb5-8842-e50c7a876e8e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.088428] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Created folder: Instances in parent group-v249000. [ 954.088722] env[62692]: DEBUG oslo.service.loopingcall [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 954.088934] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 954.089188] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f465dc67-177e-4dee-b2d7-0f0b77c42b16 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.112553] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 954.112553] env[62692]: value = "task-1141467" [ 954.112553] env[62692]: _type = "Task" [ 954.112553] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.124112] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141467, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.329399] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0fda484a-558a-41cf-b7ea-4d79bff27960 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "a668a569-a90e-4e59-8d2e-b0225745b500" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.329911] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0fda484a-558a-41cf-b7ea-4d79bff27960 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "a668a569-a90e-4e59-8d2e-b0225745b500" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.329911] env[62692]: DEBUG nova.compute.manager [None req-0fda484a-558a-41cf-b7ea-4d79bff27960 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 954.331031] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dfe360a-4e1a-438e-bed8-7516bc39d480 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.340279] env[62692]: DEBUG nova.compute.manager [None req-0fda484a-558a-41cf-b7ea-4d79bff27960 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62692) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 954.340749] env[62692]: DEBUG nova.objects.instance [None req-0fda484a-558a-41cf-b7ea-4d79bff27960 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lazy-loading 'flavor' on Instance uuid a668a569-a90e-4e59-8d2e-b0225745b500 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 954.353107] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141463, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.518649] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b429c489-fb48-4cd2-89e9-11930cbc6d4c tempest-ServersTestMultiNic-1868917293 tempest-ServersTestMultiNic-1868917293-project-member] Lock "9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.453s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.535296] env[62692]: DEBUG oslo_vmware.api [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141464, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.101567} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.535583] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 954.536409] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95802ec7-4d41-4b9e-891e-f18432a2136b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.563886] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 46bce865-188b-487c-a73f-cf524059c797/46bce865-188b-487c-a73f-cf524059c797.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 954.566401] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-566bf962-c498-4267-b556-9be0a13c0814 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.590659] env[62692]: DEBUG oslo_vmware.api [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 954.590659] env[62692]: value = "task-1141468" [ 954.590659] env[62692]: _type = "Task" [ 954.590659] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.602254] env[62692]: DEBUG oslo_vmware.api [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141468, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.624698] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141467, 'name': CreateVM_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.852870] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fda484a-558a-41cf-b7ea-4d79bff27960 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 954.852870] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a19f9714-98c5-4ef4-ba8f-8ccccb759c6d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.859774] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141463, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.685061} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.860546] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 5a08fa8a-f9fe-4879-bb7b-baa04097df6c/5a08fa8a-f9fe-4879-bb7b-baa04097df6c.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 954.860787] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 954.861287] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ff62c4a7-98e9-4bdf-9ec5-8f3a8eaf7bad {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.868948] env[62692]: DEBUG oslo_vmware.api [None req-0fda484a-558a-41cf-b7ea-4d79bff27960 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 954.868948] env[62692]: value = "task-1141469" [ 954.868948] env[62692]: _type = "Task" [ 954.868948] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.873702] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 954.873702] env[62692]: value = "task-1141470" [ 954.873702] env[62692]: _type = "Task" [ 954.873702] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.881382] env[62692]: DEBUG oslo_vmware.api [None req-0fda484a-558a-41cf-b7ea-4d79bff27960 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141469, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.890192] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141470, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.938912] env[62692]: DEBUG nova.network.neutron [req-335437f6-05d7-4d00-a49c-d52d8e53a89e req-b2e9e1e6-6547-4b78-9c8f-75126f2295f7 service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Updated VIF entry in instance network info cache for port 3fa9bd6e-092d-4e9f-a4ab-5113affb920b. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 954.939350] env[62692]: DEBUG nova.network.neutron [req-335437f6-05d7-4d00-a49c-d52d8e53a89e req-b2e9e1e6-6547-4b78-9c8f-75126f2295f7 service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Updating instance_info_cache with network_info: [{"id": "3fa9bd6e-092d-4e9f-a4ab-5113affb920b", "address": "fa:16:3e:a2:35:15", "network": {"id": "bbdf7bb9-6b54-4e19-91c6-cb0650e1121b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-2011930120-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "4b111e98b02a401384b23e363372ed52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5", "external-id": "nsx-vlan-transportzone-685", "segmentation_id": 685, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3fa9bd6e-09", "ovs_interfaceid": "3fa9bd6e-092d-4e9f-a4ab-5113affb920b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.953304] env[62692]: DEBUG nova.compute.manager [req-36d3448a-a6a1-417c-ad4c-ebc2ecc3b113 req-9d493215-7c3b-457f-a327-90dc8c993fec service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Received event network-vif-plugged-2deb4259-2f64-4f7c-9289-eba76375335a {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 954.953631] env[62692]: DEBUG oslo_concurrency.lockutils [req-36d3448a-a6a1-417c-ad4c-ebc2ecc3b113 req-9d493215-7c3b-457f-a327-90dc8c993fec service nova] Acquiring lock "668fb0ea-7eed-4198-943c-8f916ec11368-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.953733] env[62692]: DEBUG oslo_concurrency.lockutils [req-36d3448a-a6a1-417c-ad4c-ebc2ecc3b113 req-9d493215-7c3b-457f-a327-90dc8c993fec service nova] Lock "668fb0ea-7eed-4198-943c-8f916ec11368-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.953904] env[62692]: DEBUG oslo_concurrency.lockutils [req-36d3448a-a6a1-417c-ad4c-ebc2ecc3b113 req-9d493215-7c3b-457f-a327-90dc8c993fec service nova] Lock "668fb0ea-7eed-4198-943c-8f916ec11368-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.954092] env[62692]: DEBUG nova.compute.manager [req-36d3448a-a6a1-417c-ad4c-ebc2ecc3b113 req-9d493215-7c3b-457f-a327-90dc8c993fec service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] No waiting events found dispatching network-vif-plugged-2deb4259-2f64-4f7c-9289-eba76375335a {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 954.954265] env[62692]: WARNING nova.compute.manager [req-36d3448a-a6a1-417c-ad4c-ebc2ecc3b113 req-9d493215-7c3b-457f-a327-90dc8c993fec service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Received unexpected event network-vif-plugged-2deb4259-2f64-4f7c-9289-eba76375335a for instance with vm_state building and task_state spawning. [ 954.972229] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89f59000-5380-40df-8e9f-e253466d5ddb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.980521] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5020a2e-fc90-44bf-9451-54edcc586524 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.013665] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e56b379-8fa6-4d66-b63d-c83f10ce4265 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Volume attach. Driver type: vmdk {{(pid=62692) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 955.014100] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e56b379-8fa6-4d66-b63d-c83f10ce4265 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-248997', 'volume_id': '50c399be-767e-458d-bf5a-9c2395ab8e24', 'name': 'volume-50c399be-767e-458d-bf5a-9c2395ab8e24', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19', 'attached_at': '', 'detached_at': '', 'volume_id': '50c399be-767e-458d-bf5a-9c2395ab8e24', 'serial': '50c399be-767e-458d-bf5a-9c2395ab8e24'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 955.014941] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1abd3e17-8372-4cea-8b07-e4d46064a4e0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.018571] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1af6c55c-ec54-4d6c-8372-7bdacc89e285 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.040556] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ac40809-4147-46be-93fe-1f949f1fe114 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.045326] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd882393-0582-4d0a-be61-87d2bb8f10e1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.058828] env[62692]: DEBUG nova.compute.provider_tree [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 955.080240] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e56b379-8fa6-4d66-b63d-c83f10ce4265 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] volume-50c399be-767e-458d-bf5a-9c2395ab8e24/volume-50c399be-767e-458d-bf5a-9c2395ab8e24.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 955.082436] env[62692]: DEBUG nova.scheduler.client.report [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 955.085215] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-607f15f0-216a-4a2c-9eac-6858ca61baa7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.108822] env[62692]: DEBUG oslo_vmware.api [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141468, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.110335] env[62692]: DEBUG oslo_vmware.api [None req-5e56b379-8fa6-4d66-b63d-c83f10ce4265 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 955.110335] env[62692]: value = "task-1141471" [ 955.110335] env[62692]: _type = "Task" [ 955.110335] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.122804] env[62692]: DEBUG oslo_vmware.api [None req-5e56b379-8fa6-4d66-b63d-c83f10ce4265 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141471, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.127222] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141467, 'name': CreateVM_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.384356] env[62692]: DEBUG oslo_vmware.api [None req-0fda484a-558a-41cf-b7ea-4d79bff27960 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141469, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.391688] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141470, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072373} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.392225] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 955.393534] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48862f3e-303c-4eca-bb08-343c17cfbd9b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.428388] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Reconfiguring VM instance instance-00000026 to attach disk [datastore2] 5a08fa8a-f9fe-4879-bb7b-baa04097df6c/5a08fa8a-f9fe-4879-bb7b-baa04097df6c.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 955.429136] env[62692]: DEBUG nova.network.neutron [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Successfully updated port: 2deb4259-2f64-4f7c-9289-eba76375335a {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 955.430568] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-44fcabb9-e13b-414b-a5f0-5b535b454bcc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.451013] env[62692]: DEBUG oslo_concurrency.lockutils [req-335437f6-05d7-4d00-a49c-d52d8e53a89e req-b2e9e1e6-6547-4b78-9c8f-75126f2295f7 service nova] Releasing lock "refresh_cache-084ce29c-302a-45cc-b36f-0143eccee298" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.461222] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 955.461222] env[62692]: value = "task-1141472" [ 955.461222] env[62692]: _type = "Task" [ 955.461222] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.472549] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141472, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.600880] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.624s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.604408] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.016s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.606852] env[62692]: INFO nova.compute.claims [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 955.630762] env[62692]: INFO nova.scheduler.client.report [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Deleted allocations for instance f44e2ee2-6321-4237-9644-a19b5e90ae16 [ 955.652881] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141467, 'name': CreateVM_Task, 'duration_secs': 1.463784} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.653519] env[62692]: DEBUG oslo_vmware.api [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141468, 'name': ReconfigVM_Task, 'duration_secs': 1.041997} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.654025] env[62692]: DEBUG oslo_vmware.api [None req-5e56b379-8fa6-4d66-b63d-c83f10ce4265 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141471, 'name': ReconfigVM_Task, 'duration_secs': 0.422392} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.659694] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 955.659997] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 46bce865-188b-487c-a73f-cf524059c797/46bce865-188b-487c-a73f-cf524059c797.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 955.661398] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e56b379-8fa6-4d66-b63d-c83f10ce4265 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Reconfigured VM instance instance-00000040 to attach disk [datastore2] volume-50c399be-767e-458d-bf5a-9c2395ab8e24/volume-50c399be-767e-458d-bf5a-9c2395ab8e24.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 955.670883] env[62692]: DEBUG oslo_concurrency.lockutils [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.671096] env[62692]: DEBUG oslo_concurrency.lockutils [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.671502] env[62692]: DEBUG oslo_concurrency.lockutils [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 955.671637] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f7f42df6-5402-468d-b3f1-bc8b2b40e74d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.676018] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a8cad62f-ea53-455f-95a4-f61fd287350a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.683999] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0580db9e-3d3b-4126-966c-0fce42c85ce3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.691703] env[62692]: DEBUG oslo_vmware.api [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Waiting for the task: (returnval){ [ 955.691703] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52f00b4a-1add-ecf0-baa1-e22ff94e85c0" [ 955.691703] env[62692]: _type = "Task" [ 955.691703] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.694251] env[62692]: DEBUG oslo_vmware.api [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 955.694251] env[62692]: value = "task-1141473" [ 955.694251] env[62692]: _type = "Task" [ 955.694251] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.698273] env[62692]: DEBUG oslo_vmware.api [None req-5e56b379-8fa6-4d66-b63d-c83f10ce4265 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 955.698273] env[62692]: value = "task-1141474" [ 955.698273] env[62692]: _type = "Task" [ 955.698273] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.709738] env[62692]: DEBUG oslo_vmware.api [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52f00b4a-1add-ecf0-baa1-e22ff94e85c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.716359] env[62692]: DEBUG oslo_vmware.api [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141473, 'name': Rename_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.720540] env[62692]: DEBUG oslo_vmware.api [None req-5e56b379-8fa6-4d66-b63d-c83f10ce4265 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141474, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.807615] env[62692]: DEBUG nova.compute.manager [req-bec03635-e60d-48ed-99e3-319090f098b5 req-2093f07c-8caf-41f6-b373-407f17d0f9e7 service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Received event network-changed-2deb4259-2f64-4f7c-9289-eba76375335a {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 955.807912] env[62692]: DEBUG nova.compute.manager [req-bec03635-e60d-48ed-99e3-319090f098b5 req-2093f07c-8caf-41f6-b373-407f17d0f9e7 service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Refreshing instance network info cache due to event network-changed-2deb4259-2f64-4f7c-9289-eba76375335a. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 955.808436] env[62692]: DEBUG oslo_concurrency.lockutils [req-bec03635-e60d-48ed-99e3-319090f098b5 req-2093f07c-8caf-41f6-b373-407f17d0f9e7 service nova] Acquiring lock "refresh_cache-668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.808626] env[62692]: DEBUG oslo_concurrency.lockutils [req-bec03635-e60d-48ed-99e3-319090f098b5 req-2093f07c-8caf-41f6-b373-407f17d0f9e7 service nova] Acquired lock "refresh_cache-668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.808839] env[62692]: DEBUG nova.network.neutron [req-bec03635-e60d-48ed-99e3-319090f098b5 req-2093f07c-8caf-41f6-b373-407f17d0f9e7 service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Refreshing network info cache for port 2deb4259-2f64-4f7c-9289-eba76375335a {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 955.882361] env[62692]: DEBUG oslo_vmware.api [None req-0fda484a-558a-41cf-b7ea-4d79bff27960 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141469, 'name': PowerOffVM_Task, 'duration_secs': 0.56837} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.882776] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fda484a-558a-41cf-b7ea-4d79bff27960 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 955.883090] env[62692]: DEBUG nova.compute.manager [None req-0fda484a-558a-41cf-b7ea-4d79bff27960 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 955.884718] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8bd5976-1d54-4f8f-822e-94425b0c919e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.953756] env[62692]: DEBUG oslo_concurrency.lockutils [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "refresh_cache-668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.972857] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141472, 'name': ReconfigVM_Task, 'duration_secs': 0.479864} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.973176] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Reconfigured VM instance instance-00000026 to attach disk [datastore2] 5a08fa8a-f9fe-4879-bb7b-baa04097df6c/5a08fa8a-f9fe-4879-bb7b-baa04097df6c.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 955.973859] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f9fee939-3795-4460-a344-aa571b24082f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.981034] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 955.981034] env[62692]: value = "task-1141475" [ 955.981034] env[62692]: _type = "Task" [ 955.981034] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.993949] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141475, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.177177] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cfe406aa-4174-4494-9e88-25f6229b61f3 tempest-ServersTestBootFromVolume-933519659 tempest-ServersTestBootFromVolume-933519659-project-member] Lock "f44e2ee2-6321-4237-9644-a19b5e90ae16" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.446s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.207076] env[62692]: DEBUG oslo_vmware.api [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52f00b4a-1add-ecf0-baa1-e22ff94e85c0, 'name': SearchDatastore_Task, 'duration_secs': 0.016192} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.210221] env[62692]: DEBUG oslo_concurrency.lockutils [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.210221] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 956.210221] env[62692]: DEBUG oslo_concurrency.lockutils [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.210476] env[62692]: DEBUG oslo_concurrency.lockutils [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.210476] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 956.211047] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5619594c-397e-498d-8a28-922702081c9e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.216516] env[62692]: DEBUG oslo_vmware.api [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141473, 'name': Rename_Task, 'duration_secs': 0.242875} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.217133] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 956.217393] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a7bbcfc0-994a-4660-93c3-e439c7aade64 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.224019] env[62692]: DEBUG oslo_vmware.api [None req-5e56b379-8fa6-4d66-b63d-c83f10ce4265 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141474, 'name': ReconfigVM_Task, 'duration_secs': 0.195112} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.224019] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e56b379-8fa6-4d66-b63d-c83f10ce4265 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-248997', 'volume_id': '50c399be-767e-458d-bf5a-9c2395ab8e24', 'name': 'volume-50c399be-767e-458d-bf5a-9c2395ab8e24', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19', 'attached_at': '', 'detached_at': '', 'volume_id': '50c399be-767e-458d-bf5a-9c2395ab8e24', 'serial': '50c399be-767e-458d-bf5a-9c2395ab8e24'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 956.225471] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 956.225643] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 956.226567] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9634b6f9-114b-425b-a455-3713bd90415f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.230053] env[62692]: DEBUG oslo_vmware.api [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 956.230053] env[62692]: value = "task-1141476" [ 956.230053] env[62692]: _type = "Task" [ 956.230053] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.235573] env[62692]: DEBUG oslo_vmware.api [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Waiting for the task: (returnval){ [ 956.235573] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52a25c06-88d2-2eeb-f3e1-d4e203a8a2c2" [ 956.235573] env[62692]: _type = "Task" [ 956.235573] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.244143] env[62692]: DEBUG oslo_vmware.api [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141476, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.249088] env[62692]: DEBUG oslo_vmware.api [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52a25c06-88d2-2eeb-f3e1-d4e203a8a2c2, 'name': SearchDatastore_Task, 'duration_secs': 0.010187} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.249951] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eaab5419-f500-4bbd-9ce6-c9c5725a66dc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.255677] env[62692]: DEBUG oslo_vmware.api [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Waiting for the task: (returnval){ [ 956.255677] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]527537aa-d257-c099-3528-182d0c9b0e88" [ 956.255677] env[62692]: _type = "Task" [ 956.255677] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.265314] env[62692]: DEBUG oslo_vmware.api [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]527537aa-d257-c099-3528-182d0c9b0e88, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.385809] env[62692]: DEBUG nova.network.neutron [req-bec03635-e60d-48ed-99e3-319090f098b5 req-2093f07c-8caf-41f6-b373-407f17d0f9e7 service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 956.403615] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0fda484a-558a-41cf-b7ea-4d79bff27960 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "a668a569-a90e-4e59-8d2e-b0225745b500" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.074s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.496876] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141475, 'name': Rename_Task, 'duration_secs': 0.173862} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.497274] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 956.497552] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c53b22e3-91fb-431f-a139-d63bb826ebfc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.509029] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 956.509029] env[62692]: value = "task-1141477" [ 956.509029] env[62692]: _type = "Task" [ 956.509029] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.515956] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141477, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.588682] env[62692]: DEBUG nova.network.neutron [req-bec03635-e60d-48ed-99e3-319090f098b5 req-2093f07c-8caf-41f6-b373-407f17d0f9e7 service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.749936] env[62692]: DEBUG oslo_vmware.api [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141476, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.768050] env[62692]: DEBUG oslo_vmware.api [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]527537aa-d257-c099-3528-182d0c9b0e88, 'name': SearchDatastore_Task, 'duration_secs': 0.010391} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.768387] env[62692]: DEBUG oslo_concurrency.lockutils [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.768653] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 084ce29c-302a-45cc-b36f-0143eccee298/084ce29c-302a-45cc-b36f-0143eccee298.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 956.768949] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0d95e56e-692e-4a94-8e04-8a66cc724beb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.780782] env[62692]: DEBUG oslo_vmware.api [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Waiting for the task: (returnval){ [ 956.780782] env[62692]: value = "task-1141478" [ 956.780782] env[62692]: _type = "Task" [ 956.780782] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.804289] env[62692]: DEBUG oslo_vmware.api [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141478, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.021797] env[62692]: DEBUG oslo_vmware.api [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141477, 'name': PowerOnVM_Task, 'duration_secs': 0.490637} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.022294] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 957.022586] env[62692]: DEBUG nova.compute.manager [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 957.023625] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c0cf523-130e-429f-a568-df38998db9aa {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.093447] env[62692]: DEBUG oslo_concurrency.lockutils [req-bec03635-e60d-48ed-99e3-319090f098b5 req-2093f07c-8caf-41f6-b373-407f17d0f9e7 service nova] Releasing lock "refresh_cache-668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.097153] env[62692]: DEBUG oslo_concurrency.lockutils [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "refresh_cache-668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.097466] env[62692]: DEBUG nova.network.neutron [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 957.181441] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d71aa1d-7852-4c51-99b5-94461ed72cae {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.194404] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1289a660-3f79-405c-8ad5-1dfc755d23f8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.233999] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fac38762-81ba-40ef-94a2-0a48aa97fc3f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.248713] env[62692]: DEBUG oslo_vmware.api [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141476, 'name': PowerOnVM_Task, 'duration_secs': 0.576474} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.251493] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 957.251905] env[62692]: INFO nova.compute.manager [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Took 8.57 seconds to spawn the instance on the hypervisor. [ 957.252184] env[62692]: DEBUG nova.compute.manager [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 957.253192] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09d59c22-5a33-4989-9728-5396341f404a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.257508] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c01d3cff-1eb2-4c4f-9949-1ce83c91ec4c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.283210] env[62692]: DEBUG nova.compute.provider_tree [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 957.291040] env[62692]: DEBUG nova.objects.instance [None req-5e56b379-8fa6-4d66-b63d-c83f10ce4265 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lazy-loading 'flavor' on Instance uuid 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 957.297323] env[62692]: DEBUG oslo_vmware.api [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141478, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.381134] env[62692]: INFO nova.compute.manager [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Rebuilding instance [ 957.429323] env[62692]: DEBUG nova.compute.manager [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 957.430390] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ea51a2a-4aa3-4e88-9cdb-7a96ce672198 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.547625] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.646504] env[62692]: DEBUG nova.network.neutron [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 957.796642] env[62692]: DEBUG nova.scheduler.client.report [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 957.817252] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5e56b379-8fa6-4d66-b63d-c83f10ce4265 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.448s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.819393] env[62692]: INFO nova.compute.manager [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Took 37.61 seconds to build instance. [ 957.828229] env[62692]: DEBUG oslo_vmware.api [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141478, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.6109} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.828229] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 084ce29c-302a-45cc-b36f-0143eccee298/084ce29c-302a-45cc-b36f-0143eccee298.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 957.828229] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 957.828229] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-524899e6-7679-40a4-8143-1e0d86df0334 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.834940] env[62692]: DEBUG oslo_vmware.api [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Waiting for the task: (returnval){ [ 957.834940] env[62692]: value = "task-1141479" [ 957.834940] env[62692]: _type = "Task" [ 957.834940] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.846474] env[62692]: DEBUG oslo_vmware.api [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141479, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.948872] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 957.949191] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-98eaa07d-8a18-4da0-9a3c-2957e3dec8c6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.958889] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 957.958889] env[62692]: value = "task-1141480" [ 957.958889] env[62692]: _type = "Task" [ 957.958889] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.974694] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] VM already powered off {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 957.974694] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 957.974694] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ee4197f-7489-4af2-ae84-40a6ce49444b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.983329] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 957.984553] env[62692]: DEBUG nova.network.neutron [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Updating instance_info_cache with network_info: [{"id": "2deb4259-2f64-4f7c-9289-eba76375335a", "address": "fa:16:3e:bd:3f:b3", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2deb4259-2f", "ovs_interfaceid": "2deb4259-2f64-4f7c-9289-eba76375335a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.986624] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8dd09ebe-f811-4b4d-9044-1917ad0b0ea4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.082816] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 958.083075] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 958.084266] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Deleting the datastore file [datastore1] a668a569-a90e-4e59-8d2e-b0225745b500 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 958.084266] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4d060edd-943d-4b61-9e1c-b6f8fab88f9f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.093686] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 958.093686] env[62692]: value = "task-1141482" [ 958.093686] env[62692]: _type = "Task" [ 958.093686] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.102460] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141482, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.154197] env[62692]: INFO nova.compute.manager [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Rescuing [ 958.154320] env[62692]: DEBUG oslo_concurrency.lockutils [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "refresh_cache-3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.154437] env[62692]: DEBUG oslo_concurrency.lockutils [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquired lock "refresh_cache-3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.154606] env[62692]: DEBUG nova.network.neutron [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 958.322249] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.718s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.323237] env[62692]: DEBUG nova.compute.manager [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 958.330040] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9ef694f2-4b6f-467c-b79a-b9bcc6435eea tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "46bce865-188b-487c-a73f-cf524059c797" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.124s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.330745] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 25.816s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.330820] env[62692]: DEBUG nova.objects.instance [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fac426b6-747a-455e-85a1-44e71b432180] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62692) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 958.345881] env[62692]: DEBUG oslo_vmware.api [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141479, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.137654} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.347163] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 958.347470] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34be3d7c-8d8c-42b3-b38c-fae3af950c16 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.372113] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] 084ce29c-302a-45cc-b36f-0143eccee298/084ce29c-302a-45cc-b36f-0143eccee298.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 958.373142] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-30ea42a8-df22-4e98-ba58-f8b3fc66a601 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.398393] env[62692]: DEBUG oslo_vmware.api [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Waiting for the task: (returnval){ [ 958.398393] env[62692]: value = "task-1141483" [ 958.398393] env[62692]: _type = "Task" [ 958.398393] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.407439] env[62692]: DEBUG oslo_vmware.api [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141483, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.489559] env[62692]: DEBUG oslo_concurrency.lockutils [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "refresh_cache-668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.490088] env[62692]: DEBUG nova.compute.manager [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Instance network_info: |[{"id": "2deb4259-2f64-4f7c-9289-eba76375335a", "address": "fa:16:3e:bd:3f:b3", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2deb4259-2f", "ovs_interfaceid": "2deb4259-2f64-4f7c-9289-eba76375335a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 958.491720] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bd:3f:b3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6a31b224-77d7-48a4-af87-312758611b8e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2deb4259-2f64-4f7c-9289-eba76375335a', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 958.501473] env[62692]: DEBUG oslo.service.loopingcall [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 958.501611] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 958.502165] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fac13036-6208-4f9e-b674-9b8c590875a4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.523846] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 958.523846] env[62692]: value = "task-1141484" [ 958.523846] env[62692]: _type = "Task" [ 958.523846] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.533083] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141484, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.603177] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141482, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.19896} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.603472] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 958.603669] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 958.603863] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 958.832880] env[62692]: DEBUG nova.compute.utils [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 958.839626] env[62692]: DEBUG nova.compute.manager [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 958.839906] env[62692]: DEBUG nova.network.neutron [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 958.910616] env[62692]: DEBUG oslo_vmware.api [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141483, 'name': ReconfigVM_Task, 'duration_secs': 0.431111} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.911339] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Reconfigured VM instance instance-0000004e to attach disk [datastore2] 084ce29c-302a-45cc-b36f-0143eccee298/084ce29c-302a-45cc-b36f-0143eccee298.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 958.912152] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-42bc66d3-9df4-4607-97d6-038cb720c5b6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.920340] env[62692]: DEBUG oslo_vmware.api [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Waiting for the task: (returnval){ [ 958.920340] env[62692]: value = "task-1141485" [ 958.920340] env[62692]: _type = "Task" [ 958.920340] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.935647] env[62692]: DEBUG oslo_vmware.api [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141485, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.938452] env[62692]: DEBUG nova.policy [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b0c1314a9aed42e5b54a33de96f3db1f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '71b00c566cea4b21811ea1e1ef5ba0f3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 959.040216] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141484, 'name': CreateVM_Task, 'duration_secs': 0.413833} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.045022] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 959.047199] env[62692]: DEBUG oslo_concurrency.lockutils [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.048075] env[62692]: DEBUG oslo_concurrency.lockutils [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.048934] env[62692]: DEBUG oslo_concurrency.lockutils [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 959.050330] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a6929e0-70d5-4250-b148-b147f953c687 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.061240] env[62692]: DEBUG oslo_vmware.api [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 959.061240] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]524b5743-1159-3fc3-54f8-de5a905a490d" [ 959.061240] env[62692]: _type = "Task" [ 959.061240] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.071213] env[62692]: DEBUG oslo_vmware.api [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]524b5743-1159-3fc3-54f8-de5a905a490d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.123040] env[62692]: DEBUG nova.network.neutron [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Updating instance_info_cache with network_info: [{"id": "2c7422d2-dfef-4010-a99a-60c510c54a3c", "address": "fa:16:3e:71:16:ae", "network": {"id": "17e72c3b-603f-4305-96ac-89a9b4c9c7e9", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-123501993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7db5465cb5db465b92dfa570572b6e80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c7422d2-df", "ovs_interfaceid": "2c7422d2-dfef-4010-a99a-60c510c54a3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.341606] env[62692]: DEBUG nova.compute.manager [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 959.348033] env[62692]: INFO nova.compute.manager [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Rebuilding instance [ 959.351439] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7fbb557e-0ae4-4fee-a83f-e5e03edf05c6 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.021s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.352464] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.918s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.352718] env[62692]: DEBUG nova.objects.instance [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lazy-loading 'resources' on Instance uuid 3fd2959b-80cb-470c-8c2a-40b7630458ed {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.416525] env[62692]: DEBUG nova.compute.manager [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 959.416987] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6fc9a83-dcb2-422b-9c5d-e58b5710f11f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.435812] env[62692]: DEBUG oslo_vmware.api [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141485, 'name': Rename_Task, 'duration_secs': 0.301418} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.436355] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 959.440022] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-718beee9-c375-4e26-a376-a7a731aed3be {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.444450] env[62692]: DEBUG nova.network.neutron [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Successfully created port: 1b3f2a08-aff4-4b7e-8b44-bc9033d93162 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 959.449886] env[62692]: DEBUG oslo_vmware.api [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Waiting for the task: (returnval){ [ 959.449886] env[62692]: value = "task-1141486" [ 959.449886] env[62692]: _type = "Task" [ 959.449886] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.461330] env[62692]: DEBUG oslo_vmware.api [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141486, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.551018] env[62692]: DEBUG oslo_concurrency.lockutils [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.551465] env[62692]: DEBUG oslo_concurrency.lockutils [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.551562] env[62692]: DEBUG oslo_concurrency.lockutils [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.551723] env[62692]: DEBUG oslo_concurrency.lockutils [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.551900] env[62692]: DEBUG oslo_concurrency.lockutils [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.554332] env[62692]: INFO nova.compute.manager [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Terminating instance [ 959.556343] env[62692]: DEBUG nova.compute.manager [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 959.556546] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 959.557423] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1178d09b-4889-47db-ba49-d6c88cac7c44 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.566434] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 959.567091] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-91ae6768-2204-4a67-8069-ac01bb3af6da {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.575287] env[62692]: DEBUG oslo_vmware.api [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]524b5743-1159-3fc3-54f8-de5a905a490d, 'name': SearchDatastore_Task, 'duration_secs': 0.01433} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.578608] env[62692]: DEBUG oslo_concurrency.lockutils [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.578725] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 959.578982] env[62692]: DEBUG oslo_concurrency.lockutils [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.579189] env[62692]: DEBUG oslo_concurrency.lockutils [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.579465] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 959.579851] env[62692]: DEBUG oslo_vmware.api [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 959.579851] env[62692]: value = "task-1141487" [ 959.579851] env[62692]: _type = "Task" [ 959.579851] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.580107] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c5ae0303-4464-45bf-8481-656c6c52c3ac {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.593432] env[62692]: DEBUG oslo_vmware.api [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141487, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.595611] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 959.595758] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 959.596546] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-997706d4-c892-4c43-af64-6ece2f88446b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.603410] env[62692]: DEBUG oslo_vmware.api [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 959.603410] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]522b9607-73df-84e9-feeb-3db5904044c2" [ 959.603410] env[62692]: _type = "Task" [ 959.603410] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.619086] env[62692]: DEBUG oslo_vmware.api [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]522b9607-73df-84e9-feeb-3db5904044c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.623043] env[62692]: DEBUG oslo_concurrency.lockutils [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Releasing lock "refresh_cache-3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.646538] env[62692]: DEBUG nova.virt.hardware [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 959.646911] env[62692]: DEBUG nova.virt.hardware [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 959.647336] env[62692]: DEBUG nova.virt.hardware [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 959.647609] env[62692]: DEBUG nova.virt.hardware [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 959.647829] env[62692]: DEBUG nova.virt.hardware [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 959.648210] env[62692]: DEBUG nova.virt.hardware [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 959.648517] env[62692]: DEBUG nova.virt.hardware [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 959.648704] env[62692]: DEBUG nova.virt.hardware [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 959.649109] env[62692]: DEBUG nova.virt.hardware [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 959.649338] env[62692]: DEBUG nova.virt.hardware [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 959.649542] env[62692]: DEBUG nova.virt.hardware [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 959.652132] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-811997fb-7064-4b52-917f-cd18a00f2a30 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.674023] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a76e7760-7db2-4bbc-ae15-893379796e6e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.687780] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a6:f1:8d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f1187f72-32ea-484c-9bde-f40c9a5a3a03', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 959.698009] env[62692]: DEBUG oslo.service.loopingcall [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 959.699558] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 959.700277] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2db19c49-0cf1-4a61-976f-bfad882bcedc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.728056] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 959.728056] env[62692]: value = "task-1141488" [ 959.728056] env[62692]: _type = "Task" [ 959.728056] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.745620] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141488, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.940493] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 959.940493] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-20a0ea11-b0e9-4ad6-b6f1-0cdac1869032 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.953022] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 959.953022] env[62692]: value = "task-1141489" [ 959.953022] env[62692]: _type = "Task" [ 959.953022] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.978499] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141489, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.978813] env[62692]: DEBUG oslo_vmware.api [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141486, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.098584] env[62692]: DEBUG oslo_vmware.api [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141487, 'name': PowerOffVM_Task, 'duration_secs': 0.227749} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.098928] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 960.099184] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 960.099497] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-418ab895-8ca0-4216-a1a8-ae6912c5fa55 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.116660] env[62692]: DEBUG oslo_vmware.api [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]522b9607-73df-84e9-feeb-3db5904044c2, 'name': SearchDatastore_Task, 'duration_secs': 0.0099} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.117446] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7159db9-92f3-4f0b-8a42-665328144888 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.129386] env[62692]: DEBUG oslo_vmware.api [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 960.129386] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52f73dfa-b9ec-cc31-48c8-e73d8ffda6da" [ 960.129386] env[62692]: _type = "Task" [ 960.129386] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.138298] env[62692]: DEBUG oslo_vmware.api [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52f73dfa-b9ec-cc31-48c8-e73d8ffda6da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.179556] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 960.179821] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-37ee621b-a145-4383-9251-6c28e4f8fcf1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.191999] env[62692]: DEBUG oslo_vmware.api [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 960.191999] env[62692]: value = "task-1141491" [ 960.191999] env[62692]: _type = "Task" [ 960.191999] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.194710] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 960.194941] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 960.195200] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Deleting the datastore file [datastore1] 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 960.196065] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1277f2d5-4cc0-49a0-adb2-9fa4321a11bb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.203686] env[62692]: DEBUG oslo_vmware.api [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141491, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.216251] env[62692]: DEBUG oslo_vmware.api [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 960.216251] env[62692]: value = "task-1141492" [ 960.216251] env[62692]: _type = "Task" [ 960.216251] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.226563] env[62692]: DEBUG oslo_vmware.api [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141492, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.240644] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141488, 'name': CreateVM_Task, 'duration_secs': 0.428316} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.243854] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 960.245999] env[62692]: DEBUG oslo_concurrency.lockutils [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.245999] env[62692]: DEBUG oslo_concurrency.lockutils [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.245999] env[62692]: DEBUG oslo_concurrency.lockutils [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 960.245999] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f46ffdd0-df4d-4087-b84a-0eca7da9e99c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.251931] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 960.251931] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52bbb2ec-bf37-dcf7-4ebf-5a4880b5cf7f" [ 960.251931] env[62692]: _type = "Task" [ 960.251931] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.265463] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52bbb2ec-bf37-dcf7-4ebf-5a4880b5cf7f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.353905] env[62692]: DEBUG nova.compute.manager [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 960.384487] env[62692]: DEBUG nova.virt.hardware [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 960.384832] env[62692]: DEBUG nova.virt.hardware [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 960.385117] env[62692]: DEBUG nova.virt.hardware [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 960.385445] env[62692]: DEBUG nova.virt.hardware [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 960.385599] env[62692]: DEBUG nova.virt.hardware [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 960.385761] env[62692]: DEBUG nova.virt.hardware [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 960.385981] env[62692]: DEBUG nova.virt.hardware [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 960.386165] env[62692]: DEBUG nova.virt.hardware [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 960.386345] env[62692]: DEBUG nova.virt.hardware [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 960.386520] env[62692]: DEBUG nova.virt.hardware [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 960.386699] env[62692]: DEBUG nova.virt.hardware [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 960.387605] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fffc69dd-36f4-4ada-8d73-1e0e3b1b024d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.391982] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-918613dc-4e5f-41fe-863e-d0ea2970dee6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.402080] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-befcbc23-e6c9-41db-a452-ce2f353ef253 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.406440] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad0db0ad-9804-4fd6-ade1-5670fed89eda {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.446663] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d05aa8cd-d812-4ff5-9cef-783bdce8b425 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.459028] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b89779b-da1c-4e93-9295-f3688fb548f2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.471070] env[62692]: DEBUG oslo_vmware.api [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141486, 'name': PowerOnVM_Task, 'duration_secs': 0.652952} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.481479] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 960.481754] env[62692]: INFO nova.compute.manager [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Took 9.15 seconds to spawn the instance on the hypervisor. [ 960.481955] env[62692]: DEBUG nova.compute.manager [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 960.482465] env[62692]: DEBUG nova.compute.provider_tree [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 960.483711] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141489, 'name': PowerOffVM_Task, 'duration_secs': 0.27237} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.484458] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a728fd5-f80f-4162-a4df-e64dcc5841c5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.487153] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 960.487981] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 960.489076] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4683e606-9b2c-49ea-99b0-c1942be522a2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.502641] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 960.503113] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-47f54b9d-c907-4acd-b283-d571669498b9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.582092] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 960.583751] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 960.583751] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Deleting the datastore file [datastore1] 46bce865-188b-487c-a73f-cf524059c797 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 960.583751] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-48266803-1f37-4d55-b51d-18ad1b988afe {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.591336] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 960.591336] env[62692]: value = "task-1141494" [ 960.591336] env[62692]: _type = "Task" [ 960.591336] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.600542] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141494, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.638700] env[62692]: DEBUG oslo_vmware.api [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52f73dfa-b9ec-cc31-48c8-e73d8ffda6da, 'name': SearchDatastore_Task, 'duration_secs': 0.018574} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.638972] env[62692]: DEBUG oslo_concurrency.lockutils [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.639282] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 668fb0ea-7eed-4198-943c-8f916ec11368/668fb0ea-7eed-4198-943c-8f916ec11368.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 960.639614] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3f9365df-50c4-492c-8b44-35f22cb5a3b6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.648525] env[62692]: DEBUG oslo_vmware.api [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 960.648525] env[62692]: value = "task-1141495" [ 960.648525] env[62692]: _type = "Task" [ 960.648525] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.657424] env[62692]: DEBUG oslo_vmware.api [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141495, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.701346] env[62692]: DEBUG oslo_vmware.api [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141491, 'name': PowerOffVM_Task, 'duration_secs': 0.306634} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.701753] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 960.702619] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a6950f8-0b68-4d85-aabe-142e8ee73093 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.732228] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06560d0c-91a3-4936-b9b3-09eb4f136604 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.753143] env[62692]: DEBUG oslo_vmware.api [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141492, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.174973} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.760208] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 960.760462] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 960.762973] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 960.763364] env[62692]: INFO nova.compute.manager [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Took 1.21 seconds to destroy the instance on the hypervisor. [ 960.763870] env[62692]: DEBUG oslo.service.loopingcall [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 960.764097] env[62692]: DEBUG nova.compute.manager [-] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 960.764252] env[62692]: DEBUG nova.network.neutron [-] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 960.774870] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52bbb2ec-bf37-dcf7-4ebf-5a4880b5cf7f, 'name': SearchDatastore_Task, 'duration_secs': 0.011686} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.774870] env[62692]: DEBUG oslo_concurrency.lockutils [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.774870] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 960.774870] env[62692]: DEBUG oslo_concurrency.lockutils [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.774870] env[62692]: DEBUG oslo_concurrency.lockutils [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.774870] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 960.775561] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-41cc4570-b945-4ed2-8375-94f85306aba4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.786854] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 960.787090] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 960.789648] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 960.789884] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f363d186-68e3-4e6e-bd8b-ca5963dcdd31 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.794219] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-167aa243-3477-41cc-99ae-129ae1539d53 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.799094] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 960.799094] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5255a9cc-9b42-ea58-8feb-fccf811ceef0" [ 960.799094] env[62692]: _type = "Task" [ 960.799094] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.805396] env[62692]: DEBUG oslo_vmware.api [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 960.805396] env[62692]: value = "task-1141496" [ 960.805396] env[62692]: _type = "Task" [ 960.805396] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.813752] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5255a9cc-9b42-ea58-8feb-fccf811ceef0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.819534] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] VM already powered off {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 960.819742] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 960.819983] env[62692]: DEBUG oslo_concurrency.lockutils [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.820178] env[62692]: DEBUG oslo_concurrency.lockutils [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.820346] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 960.820597] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e0cc1c4a-db05-4d49-a73b-49957131bbef {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.832920] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 960.833134] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 960.833998] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7532c59a-0c4f-4f0c-8185-ab3f4eb421f6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.840108] env[62692]: DEBUG oslo_vmware.api [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 960.840108] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5246fce3-35a0-a7b0-755d-49677269174c" [ 960.840108] env[62692]: _type = "Task" [ 960.840108] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.848750] env[62692]: DEBUG oslo_vmware.api [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5246fce3-35a0-a7b0-755d-49677269174c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.990441] env[62692]: DEBUG nova.scheduler.client.report [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 961.011145] env[62692]: INFO nova.compute.manager [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Took 38.29 seconds to build instance. [ 961.071817] env[62692]: DEBUG nova.compute.manager [req-b3db91d6-a983-4624-8216-69dfe429b692 req-00f31efa-483b-4e6b-81fa-9114b7dd05be service nova] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Received event network-vif-deleted-0047b080-5217-4718-a134-a87f4556102b {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 961.071916] env[62692]: INFO nova.compute.manager [req-b3db91d6-a983-4624-8216-69dfe429b692 req-00f31efa-483b-4e6b-81fa-9114b7dd05be service nova] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Neutron deleted interface 0047b080-5217-4718-a134-a87f4556102b; detaching it from the instance and deleting it from the info cache [ 961.072106] env[62692]: DEBUG nova.network.neutron [req-b3db91d6-a983-4624-8216-69dfe429b692 req-00f31efa-483b-4e6b-81fa-9114b7dd05be service nova] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.105109] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141494, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.37832} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.106361] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 961.106565] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 961.106749] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 961.163458] env[62692]: DEBUG oslo_vmware.api [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141495, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.313535] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5255a9cc-9b42-ea58-8feb-fccf811ceef0, 'name': SearchDatastore_Task, 'duration_secs': 0.029139} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.314443] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a09af6d-56c1-46ef-bd8e-d604c44f7f7d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.323862] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 961.323862] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5224fafb-f6e0-404a-0277-556a0cecde22" [ 961.323862] env[62692]: _type = "Task" [ 961.323862] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.337526] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5224fafb-f6e0-404a-0277-556a0cecde22, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.353277] env[62692]: DEBUG oslo_vmware.api [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5246fce3-35a0-a7b0-755d-49677269174c, 'name': SearchDatastore_Task, 'duration_secs': 0.020483} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.354465] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-769bc472-246b-4f6b-8f9d-86d087bb4af2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.364598] env[62692]: DEBUG oslo_vmware.api [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 961.364598] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52a52fdb-4893-0bf4-f1a7-ec3211e70535" [ 961.364598] env[62692]: _type = "Task" [ 961.364598] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.376480] env[62692]: DEBUG oslo_vmware.api [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52a52fdb-4893-0bf4-f1a7-ec3211e70535, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.495977] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.143s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.500495] env[62692]: DEBUG oslo_concurrency.lockutils [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.465s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.501793] env[62692]: INFO nova.compute.claims [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 961.513795] env[62692]: DEBUG oslo_concurrency.lockutils [None req-af084158-52de-4980-8bdf-6e40c8a4a65a tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Lock "084ce29c-302a-45cc-b36f-0143eccee298" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.634s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.529808] env[62692]: INFO nova.scheduler.client.report [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Deleted allocations for instance 3fd2959b-80cb-470c-8c2a-40b7630458ed [ 961.548678] env[62692]: DEBUG nova.network.neutron [-] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.574398] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-185aacdc-8f69-4e95-9b6b-7730ad6beb3d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.584911] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49c01861-edd4-41a5-9c3e-c2e8f936cf0e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.626197] env[62692]: DEBUG nova.compute.manager [req-b3db91d6-a983-4624-8216-69dfe429b692 req-00f31efa-483b-4e6b-81fa-9114b7dd05be service nova] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Detach interface failed, port_id=0047b080-5217-4718-a134-a87f4556102b, reason: Instance 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 961.659439] env[62692]: DEBUG oslo_vmware.api [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141495, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.753759} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.660429] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 668fb0ea-7eed-4198-943c-8f916ec11368/668fb0ea-7eed-4198-943c-8f916ec11368.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 961.660824] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 961.661439] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9a7e836e-23bb-4a5a-a94a-b1ada99fdfec {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.673399] env[62692]: DEBUG oslo_vmware.api [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 961.673399] env[62692]: value = "task-1141497" [ 961.673399] env[62692]: _type = "Task" [ 961.673399] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.692965] env[62692]: DEBUG oslo_vmware.api [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141497, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.711181] env[62692]: DEBUG nova.network.neutron [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Successfully updated port: 1b3f2a08-aff4-4b7e-8b44-bc9033d93162 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 961.838085] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5224fafb-f6e0-404a-0277-556a0cecde22, 'name': SearchDatastore_Task, 'duration_secs': 0.021275} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.838085] env[62692]: DEBUG oslo_concurrency.lockutils [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.839327] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] a668a569-a90e-4e59-8d2e-b0225745b500/a668a569-a90e-4e59-8d2e-b0225745b500.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 961.839327] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e9763ba6-89cd-47b0-a52f-29ec0b71b2db {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.846287] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 961.846287] env[62692]: value = "task-1141498" [ 961.846287] env[62692]: _type = "Task" [ 961.846287] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.854809] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141498, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.875943] env[62692]: DEBUG oslo_vmware.api [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52a52fdb-4893-0bf4-f1a7-ec3211e70535, 'name': SearchDatastore_Task, 'duration_secs': 0.025174} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.876289] env[62692]: DEBUG oslo_concurrency.lockutils [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.876585] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19/81eb64fb-5a0c-47b7-9948-7dcb01b911bd-rescue.vmdk. {{(pid=62692) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 961.876881] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-264542bc-4b56-4173-8c71-12fdbc5a2464 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.885136] env[62692]: DEBUG oslo_vmware.api [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 961.885136] env[62692]: value = "task-1141499" [ 961.885136] env[62692]: _type = "Task" [ 961.885136] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.894948] env[62692]: DEBUG oslo_vmware.api [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141499, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.028064] env[62692]: INFO nova.compute.manager [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Rescuing [ 962.028371] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Acquiring lock "refresh_cache-084ce29c-302a-45cc-b36f-0143eccee298" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.028553] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Acquired lock "refresh_cache-084ce29c-302a-45cc-b36f-0143eccee298" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.028766] env[62692]: DEBUG nova.network.neutron [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 962.043659] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1885960e-d954-4f14-b584-2badd60e3966 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "3fd2959b-80cb-470c-8c2a-40b7630458ed" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.457s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.051467] env[62692]: INFO nova.compute.manager [-] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Took 1.29 seconds to deallocate network for instance. [ 962.160929] env[62692]: DEBUG nova.virt.hardware [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 962.161375] env[62692]: DEBUG nova.virt.hardware [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 962.161650] env[62692]: DEBUG nova.virt.hardware [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 962.161900] env[62692]: DEBUG nova.virt.hardware [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 962.162106] env[62692]: DEBUG nova.virt.hardware [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 962.162295] env[62692]: DEBUG nova.virt.hardware [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 962.162545] env[62692]: DEBUG nova.virt.hardware [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 962.162719] env[62692]: DEBUG nova.virt.hardware [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 962.162973] env[62692]: DEBUG nova.virt.hardware [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 962.163266] env[62692]: DEBUG nova.virt.hardware [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 962.163553] env[62692]: DEBUG nova.virt.hardware [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 962.165327] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e271468d-8ab2-4ead-be4b-cd4b1face6e7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.178896] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28df8bb6-8d01-4500-97a0-97172e8f4a5d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.191483] env[62692]: DEBUG oslo_vmware.api [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141497, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073089} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.199632] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 962.200328] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:04:13:61', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f44b2fa3-6730-4b87-8839-947eff21213f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5ff6939b-f0fd-4b3d-b0e8-f7946f80493b', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 962.208105] env[62692]: DEBUG oslo.service.loopingcall [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 962.208944] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf48aa29-b079-4343-ac0f-9da5968c6620 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.212372] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 46bce865-188b-487c-a73f-cf524059c797] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 962.212679] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d843a8b1-447a-4663-be7d-d5d58faf07ae {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.234454] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "refresh_cache-f3e8be1c-d952-49b7-86a8-d0a6c56ee096" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.234628] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquired lock "refresh_cache-f3e8be1c-d952-49b7-86a8-d0a6c56ee096" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.234781] env[62692]: DEBUG nova.network.neutron [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 962.257926] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] 668fb0ea-7eed-4198-943c-8f916ec11368/668fb0ea-7eed-4198-943c-8f916ec11368.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 962.260550] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-063f7514-b63b-4a96-8127-60e9ec442e9c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.279627] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 962.279627] env[62692]: value = "task-1141500" [ 962.279627] env[62692]: _type = "Task" [ 962.279627] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.288048] env[62692]: DEBUG oslo_vmware.api [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 962.288048] env[62692]: value = "task-1141501" [ 962.288048] env[62692]: _type = "Task" [ 962.288048] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.292324] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141500, 'name': CreateVM_Task} progress is 15%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.304544] env[62692]: DEBUG oslo_vmware.api [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141501, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.359343] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 962.359804] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141498, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.360091] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 962.397418] env[62692]: DEBUG oslo_vmware.api [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141499, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.561284] env[62692]: DEBUG oslo_concurrency.lockutils [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.791541] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141500, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.800126] env[62692]: DEBUG nova.network.neutron [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 962.807117] env[62692]: DEBUG oslo_vmware.api [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141501, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.861443] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141498, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.628836} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.861800] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] a668a569-a90e-4e59-8d2e-b0225745b500/a668a569-a90e-4e59-8d2e-b0225745b500.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 962.862029] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 962.862566] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b32263fe-fa95-459c-8a4c-16b4d22053ff {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.874623] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 962.874919] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Starting heal instance info cache {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 962.881031] env[62692]: DEBUG nova.network.neutron [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Updating instance_info_cache with network_info: [{"id": "3fa9bd6e-092d-4e9f-a4ab-5113affb920b", "address": "fa:16:3e:a2:35:15", "network": {"id": "bbdf7bb9-6b54-4e19-91c6-cb0650e1121b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-2011930120-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "4b111e98b02a401384b23e363372ed52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5", "external-id": "nsx-vlan-transportzone-685", "segmentation_id": 685, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3fa9bd6e-09", "ovs_interfaceid": "3fa9bd6e-092d-4e9f-a4ab-5113affb920b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.881031] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 962.881031] env[62692]: value = "task-1141502" [ 962.881031] env[62692]: _type = "Task" [ 962.881031] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.896868] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141502, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.911238] env[62692]: DEBUG oslo_vmware.api [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141499, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.615821} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.914493] env[62692]: INFO nova.virt.vmwareapi.ds_util [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19/81eb64fb-5a0c-47b7-9948-7dcb01b911bd-rescue.vmdk. [ 962.915836] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2376b464-e41b-4a3b-92fc-1557ab5b97fc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.946960] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19/81eb64fb-5a0c-47b7-9948-7dcb01b911bd-rescue.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 962.952929] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-efd6f3c1-59ba-4989-863f-3424b9705a01 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.977260] env[62692]: DEBUG oslo_vmware.api [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 962.977260] env[62692]: value = "task-1141503" [ 962.977260] env[62692]: _type = "Task" [ 962.977260] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.990352] env[62692]: DEBUG oslo_vmware.api [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141503, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.095397] env[62692]: DEBUG nova.network.neutron [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Updating instance_info_cache with network_info: [{"id": "1b3f2a08-aff4-4b7e-8b44-bc9033d93162", "address": "fa:16:3e:de:6b:59", "network": {"id": "e9c538a3-7723-42cb-85b7-5a1597e49cef", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1878627765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71b00c566cea4b21811ea1e1ef5ba0f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b3f2a08-af", "ovs_interfaceid": "1b3f2a08-aff4-4b7e-8b44-bc9033d93162", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.098161] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-698ffe50-fd0f-4746-a871-88bee25f62b7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.106374] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af8eacee-4211-467c-aa31-d4687cbeb995 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.138421] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a350d90-0b8e-40e3-b1f2-397ae33e66da {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.148330] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc368876-f266-4f83-bfc3-a6ce42ad1354 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.168300] env[62692]: DEBUG nova.compute.provider_tree [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.299727] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141500, 'name': CreateVM_Task, 'duration_secs': 0.576011} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.300310] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 46bce865-188b-487c-a73f-cf524059c797] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 963.301219] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.301421] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.301724] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 963.302222] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-212af7ca-8d64-4716-81c0-1aeb1133817a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.308024] env[62692]: DEBUG oslo_vmware.api [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141501, 'name': ReconfigVM_Task, 'duration_secs': 0.5417} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.308919] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Reconfigured VM instance instance-0000004f to attach disk [datastore2] 668fb0ea-7eed-4198-943c-8f916ec11368/668fb0ea-7eed-4198-943c-8f916ec11368.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 963.309604] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e1979671-2f27-4bd7-b0bc-3f43891dd21e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.312338] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 963.312338] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5298bcc9-c027-55b4-e6aa-e57ee8a0cd32" [ 963.312338] env[62692]: _type = "Task" [ 963.312338] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.318077] env[62692]: DEBUG oslo_vmware.api [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 963.318077] env[62692]: value = "task-1141504" [ 963.318077] env[62692]: _type = "Task" [ 963.318077] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.321154] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5298bcc9-c027-55b4-e6aa-e57ee8a0cd32, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.329288] env[62692]: DEBUG oslo_vmware.api [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141504, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.381678] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Releasing lock "refresh_cache-084ce29c-302a-45cc-b36f-0143eccee298" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.393831] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141502, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.113429} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.394161] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 963.394952] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9069cd35-f9e8-4472-9882-19f63aa1826f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.421637] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] a668a569-a90e-4e59-8d2e-b0225745b500/a668a569-a90e-4e59-8d2e-b0225745b500.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 963.424699] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-876421a8-9f00-4061-b7a5-d3ac1046f0b0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.447125] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 963.447125] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5218e462-8288-49e3-87f3-d4de4d96053d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.452015] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 963.452015] env[62692]: value = "task-1141505" [ 963.452015] env[62692]: _type = "Task" [ 963.452015] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.457136] env[62692]: DEBUG oslo_vmware.api [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Waiting for the task: (returnval){ [ 963.457136] env[62692]: value = "task-1141506" [ 963.457136] env[62692]: _type = "Task" [ 963.457136] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.463944] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141505, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.471988] env[62692]: DEBUG oslo_vmware.api [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141506, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.489499] env[62692]: DEBUG oslo_vmware.api [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141503, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.524812] env[62692]: DEBUG nova.compute.manager [req-4feac038-4aa0-4ed1-b83c-c763bf606337 req-c4eac781-92b4-4eaa-bd6f-730e5837adea service nova] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Received event network-vif-plugged-1b3f2a08-aff4-4b7e-8b44-bc9033d93162 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 963.524970] env[62692]: DEBUG oslo_concurrency.lockutils [req-4feac038-4aa0-4ed1-b83c-c763bf606337 req-c4eac781-92b4-4eaa-bd6f-730e5837adea service nova] Acquiring lock "f3e8be1c-d952-49b7-86a8-d0a6c56ee096-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.525285] env[62692]: DEBUG oslo_concurrency.lockutils [req-4feac038-4aa0-4ed1-b83c-c763bf606337 req-c4eac781-92b4-4eaa-bd6f-730e5837adea service nova] Lock "f3e8be1c-d952-49b7-86a8-d0a6c56ee096-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.525517] env[62692]: DEBUG oslo_concurrency.lockutils [req-4feac038-4aa0-4ed1-b83c-c763bf606337 req-c4eac781-92b4-4eaa-bd6f-730e5837adea service nova] Lock "f3e8be1c-d952-49b7-86a8-d0a6c56ee096-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.526728] env[62692]: DEBUG nova.compute.manager [req-4feac038-4aa0-4ed1-b83c-c763bf606337 req-c4eac781-92b4-4eaa-bd6f-730e5837adea service nova] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] No waiting events found dispatching network-vif-plugged-1b3f2a08-aff4-4b7e-8b44-bc9033d93162 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 963.527060] env[62692]: WARNING nova.compute.manager [req-4feac038-4aa0-4ed1-b83c-c763bf606337 req-c4eac781-92b4-4eaa-bd6f-730e5837adea service nova] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Received unexpected event network-vif-plugged-1b3f2a08-aff4-4b7e-8b44-bc9033d93162 for instance with vm_state building and task_state spawning. [ 963.527514] env[62692]: DEBUG nova.compute.manager [req-4feac038-4aa0-4ed1-b83c-c763bf606337 req-c4eac781-92b4-4eaa-bd6f-730e5837adea service nova] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Received event network-changed-1b3f2a08-aff4-4b7e-8b44-bc9033d93162 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 963.527608] env[62692]: DEBUG nova.compute.manager [req-4feac038-4aa0-4ed1-b83c-c763bf606337 req-c4eac781-92b4-4eaa-bd6f-730e5837adea service nova] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Refreshing instance network info cache due to event network-changed-1b3f2a08-aff4-4b7e-8b44-bc9033d93162. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 963.527773] env[62692]: DEBUG oslo_concurrency.lockutils [req-4feac038-4aa0-4ed1-b83c-c763bf606337 req-c4eac781-92b4-4eaa-bd6f-730e5837adea service nova] Acquiring lock "refresh_cache-f3e8be1c-d952-49b7-86a8-d0a6c56ee096" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.602161] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Releasing lock "refresh_cache-f3e8be1c-d952-49b7-86a8-d0a6c56ee096" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.602655] env[62692]: DEBUG nova.compute.manager [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Instance network_info: |[{"id": "1b3f2a08-aff4-4b7e-8b44-bc9033d93162", "address": "fa:16:3e:de:6b:59", "network": {"id": "e9c538a3-7723-42cb-85b7-5a1597e49cef", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1878627765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71b00c566cea4b21811ea1e1ef5ba0f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b3f2a08-af", "ovs_interfaceid": "1b3f2a08-aff4-4b7e-8b44-bc9033d93162", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 963.602873] env[62692]: DEBUG oslo_concurrency.lockutils [req-4feac038-4aa0-4ed1-b83c-c763bf606337 req-c4eac781-92b4-4eaa-bd6f-730e5837adea service nova] Acquired lock "refresh_cache-f3e8be1c-d952-49b7-86a8-d0a6c56ee096" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.603103] env[62692]: DEBUG nova.network.neutron [req-4feac038-4aa0-4ed1-b83c-c763bf606337 req-c4eac781-92b4-4eaa-bd6f-730e5837adea service nova] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Refreshing network info cache for port 1b3f2a08-aff4-4b7e-8b44-bc9033d93162 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 963.605367] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:de:6b:59', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dd96b39f-bd2e-48d1-85c3-577cf97f08c8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1b3f2a08-aff4-4b7e-8b44-bc9033d93162', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 963.614959] env[62692]: DEBUG oslo.service.loopingcall [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 963.616342] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 963.616597] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-91466ea8-0a40-4880-b75d-e3c06e05e66b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.643369] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 963.643369] env[62692]: value = "task-1141507" [ 963.643369] env[62692]: _type = "Task" [ 963.643369] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.660597] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141507, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.672307] env[62692]: DEBUG nova.scheduler.client.report [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 963.830026] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5298bcc9-c027-55b4-e6aa-e57ee8a0cd32, 'name': SearchDatastore_Task, 'duration_secs': 0.012258} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.833250] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.834034] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 963.834034] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.834034] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.834217] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 963.835764] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ce7136cd-2d51-411a-b793-5b9ab88403f4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.842534] env[62692]: DEBUG oslo_concurrency.lockutils [None req-fc481d96-844d-476f-93ea-68033636a9f6 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "43d5f0b7-c953-4ca8-ae9b-e45d326e008f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.842831] env[62692]: DEBUG oslo_concurrency.lockutils [None req-fc481d96-844d-476f-93ea-68033636a9f6 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "43d5f0b7-c953-4ca8-ae9b-e45d326e008f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.848417] env[62692]: DEBUG oslo_vmware.api [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141504, 'name': Rename_Task, 'duration_secs': 0.26522} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.848508] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 963.848965] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ca5e222c-4da5-4537-ac2f-bfd0d4eb2397 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.852381] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 963.852594] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 963.853342] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a5347f73-ec17-4d20-9ee9-3aa2e315431f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.857441] env[62692]: DEBUG oslo_vmware.api [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 963.857441] env[62692]: value = "task-1141508" [ 963.857441] env[62692]: _type = "Task" [ 963.857441] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.863052] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 963.863052] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]527f5bc4-68ec-9a97-bebb-a48794c87cac" [ 963.863052] env[62692]: _type = "Task" [ 963.863052] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.874016] env[62692]: DEBUG oslo_vmware.api [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141508, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.880763] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]527f5bc4-68ec-9a97-bebb-a48794c87cac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.914578] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "refresh_cache-abee88ff-5cf5-4bf1-91e0-93b19cf30046" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.914737] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquired lock "refresh_cache-abee88ff-5cf5-4bf1-91e0-93b19cf30046" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.914912] env[62692]: DEBUG nova.network.neutron [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Forcefully refreshing network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 963.964364] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141505, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.973986] env[62692]: DEBUG oslo_vmware.api [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141506, 'name': PowerOffVM_Task, 'duration_secs': 0.251028} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.974274] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 963.975073] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f2a8e5c-2392-460b-9f3f-8e8343131ae8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.997908] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e525101b-c7ac-4be8-8d9d-3d7ea040d096 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.004520] env[62692]: DEBUG oslo_vmware.api [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141503, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.048128] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 964.048927] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-015c2f1e-304a-4423-a0ad-de036c95ca4d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.064052] env[62692]: DEBUG oslo_vmware.api [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Waiting for the task: (returnval){ [ 964.064052] env[62692]: value = "task-1141509" [ 964.064052] env[62692]: _type = "Task" [ 964.064052] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.075307] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] VM already powered off {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 964.075510] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 964.075652] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.075814] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.077625] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 964.077625] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-92bfda34-b57b-4c8f-97b0-e97dbb43cf92 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.086655] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 964.086894] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 964.087798] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c3a6ceb-c017-4e02-9e15-14836952c974 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.094341] env[62692]: DEBUG oslo_vmware.api [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Waiting for the task: (returnval){ [ 964.094341] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52b1d685-ad4c-19c3-ebe3-c28df8298bb1" [ 964.094341] env[62692]: _type = "Task" [ 964.094341] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.109040] env[62692]: DEBUG oslo_vmware.api [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52b1d685-ad4c-19c3-ebe3-c28df8298bb1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.158417] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141507, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.182361] env[62692]: DEBUG oslo_concurrency.lockutils [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.681s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.182361] env[62692]: DEBUG nova.compute.manager [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 964.186314] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.057s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.186570] env[62692]: DEBUG nova.objects.instance [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lazy-loading 'resources' on Instance uuid 44699df6-5262-4700-a096-8c0e450323d2 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 964.345333] env[62692]: DEBUG nova.compute.manager [None req-fc481d96-844d-476f-93ea-68033636a9f6 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 43d5f0b7-c953-4ca8-ae9b-e45d326e008f] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 964.372289] env[62692]: DEBUG oslo_vmware.api [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141508, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.380021] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]527f5bc4-68ec-9a97-bebb-a48794c87cac, 'name': SearchDatastore_Task, 'duration_secs': 0.032013} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.380857] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-853cff2b-7327-4e88-84d0-2c56d90c544f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.390041] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 964.390041] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52ee9a77-32cb-823d-9992-67a6e75d598d" [ 964.390041] env[62692]: _type = "Task" [ 964.390041] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.399831] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52ee9a77-32cb-823d-9992-67a6e75d598d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.446030] env[62692]: DEBUG nova.network.neutron [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 964.466058] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141505, 'name': ReconfigVM_Task, 'duration_secs': 0.639309} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.466367] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Reconfigured VM instance instance-0000004c to attach disk [datastore1] a668a569-a90e-4e59-8d2e-b0225745b500/a668a569-a90e-4e59-8d2e-b0225745b500.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 964.467157] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-943283d1-2599-4532-8225-a17fda6b21c4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.471743] env[62692]: DEBUG nova.network.neutron [req-4feac038-4aa0-4ed1-b83c-c763bf606337 req-c4eac781-92b4-4eaa-bd6f-730e5837adea service nova] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Updated VIF entry in instance network info cache for port 1b3f2a08-aff4-4b7e-8b44-bc9033d93162. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 964.471743] env[62692]: DEBUG nova.network.neutron [req-4feac038-4aa0-4ed1-b83c-c763bf606337 req-c4eac781-92b4-4eaa-bd6f-730e5837adea service nova] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Updating instance_info_cache with network_info: [{"id": "1b3f2a08-aff4-4b7e-8b44-bc9033d93162", "address": "fa:16:3e:de:6b:59", "network": {"id": "e9c538a3-7723-42cb-85b7-5a1597e49cef", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1878627765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71b00c566cea4b21811ea1e1ef5ba0f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b3f2a08-af", "ovs_interfaceid": "1b3f2a08-aff4-4b7e-8b44-bc9033d93162", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.478769] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 964.478769] env[62692]: value = "task-1141510" [ 964.478769] env[62692]: _type = "Task" [ 964.478769] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.497939] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141510, 'name': Rename_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.501749] env[62692]: DEBUG oslo_vmware.api [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141503, 'name': ReconfigVM_Task, 'duration_secs': 1.244316} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.502126] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19/81eb64fb-5a0c-47b7-9948-7dcb01b911bd-rescue.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 964.503066] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b433da08-44d0-4b28-b588-67a8a76164b8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.539136] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c62a8a05-dd3e-408c-9782-7e410bd521e6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.557108] env[62692]: DEBUG oslo_vmware.api [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 964.557108] env[62692]: value = "task-1141511" [ 964.557108] env[62692]: _type = "Task" [ 964.557108] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.567101] env[62692]: DEBUG oslo_vmware.api [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141511, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.608718] env[62692]: DEBUG oslo_vmware.api [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52b1d685-ad4c-19c3-ebe3-c28df8298bb1, 'name': SearchDatastore_Task, 'duration_secs': 0.014128} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.609673] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ee1a1d9-12a1-4249-9d96-7a99c6dbaaa6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.616777] env[62692]: DEBUG oslo_vmware.api [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Waiting for the task: (returnval){ [ 964.616777] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5246c108-2a97-6bc3-85b6-6bcc891b26f4" [ 964.616777] env[62692]: _type = "Task" [ 964.616777] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.628160] env[62692]: DEBUG oslo_vmware.api [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5246c108-2a97-6bc3-85b6-6bcc891b26f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.661268] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141507, 'name': CreateVM_Task, 'duration_secs': 0.614767} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.661268] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 964.661268] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.661268] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.661412] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 964.661586] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8954bd4c-1acc-4e86-888b-be64f09d4c08 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.668096] env[62692]: DEBUG oslo_vmware.api [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 964.668096] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52d0a4de-b10c-975e-a51f-34a46ddf8f96" [ 964.668096] env[62692]: _type = "Task" [ 964.668096] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.679622] env[62692]: DEBUG oslo_vmware.api [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52d0a4de-b10c-975e-a51f-34a46ddf8f96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.690947] env[62692]: DEBUG nova.compute.utils [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 964.695092] env[62692]: DEBUG nova.compute.manager [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 964.695331] env[62692]: DEBUG nova.network.neutron [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 964.828366] env[62692]: DEBUG nova.policy [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f2e2ee26163841ddb15bb67520e98394', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e13f8888d5e84eed8247101558e29a56', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 964.849630] env[62692]: DEBUG nova.compute.manager [None req-fc481d96-844d-476f-93ea-68033636a9f6 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 43d5f0b7-c953-4ca8-ae9b-e45d326e008f] Instance disappeared before build. {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 964.873335] env[62692]: DEBUG oslo_vmware.api [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141508, 'name': PowerOnVM_Task, 'duration_secs': 0.643144} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.873669] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 964.873890] env[62692]: INFO nova.compute.manager [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Took 10.93 seconds to spawn the instance on the hypervisor. [ 964.874083] env[62692]: DEBUG nova.compute.manager [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 964.875623] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff22c787-7391-4639-bb6d-baebec6c5333 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.900516] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52ee9a77-32cb-823d-9992-67a6e75d598d, 'name': SearchDatastore_Task, 'duration_secs': 0.012998} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.904271] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.905142] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 46bce865-188b-487c-a73f-cf524059c797/46bce865-188b-487c-a73f-cf524059c797.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 964.905354] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-67e54b7b-2d8f-4e77-9cd9-352c1bcd632c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.914684] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 964.914684] env[62692]: value = "task-1141512" [ 964.914684] env[62692]: _type = "Task" [ 964.914684] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.932445] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141512, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.975927] env[62692]: DEBUG oslo_concurrency.lockutils [req-4feac038-4aa0-4ed1-b83c-c763bf606337 req-c4eac781-92b4-4eaa-bd6f-730e5837adea service nova] Releasing lock "refresh_cache-f3e8be1c-d952-49b7-86a8-d0a6c56ee096" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.989370] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141510, 'name': Rename_Task, 'duration_secs': 0.225386} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.989713] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 964.990088] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b0a02f99-7775-43c8-a533-4e77d495375d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.004818] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 965.004818] env[62692]: value = "task-1141513" [ 965.004818] env[62692]: _type = "Task" [ 965.004818] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.014614] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141513, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.071076] env[62692]: DEBUG oslo_vmware.api [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141511, 'name': ReconfigVM_Task, 'duration_secs': 0.212918} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.071076] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 965.071076] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0e507fdd-3045-466a-a5f6-55f5141a62db {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.079932] env[62692]: DEBUG oslo_vmware.api [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 965.079932] env[62692]: value = "task-1141514" [ 965.079932] env[62692]: _type = "Task" [ 965.079932] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.094861] env[62692]: DEBUG oslo_vmware.api [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141514, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.133989] env[62692]: DEBUG oslo_vmware.api [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5246c108-2a97-6bc3-85b6-6bcc891b26f4, 'name': SearchDatastore_Task, 'duration_secs': 0.011162} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.137637] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.137948] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 084ce29c-302a-45cc-b36f-0143eccee298/81eb64fb-5a0c-47b7-9948-7dcb01b911bd-rescue.vmdk. {{(pid=62692) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 965.138606] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1a9ffc5a-47f2-42a5-b959-7c834db14bfc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.151130] env[62692]: DEBUG oslo_vmware.api [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Waiting for the task: (returnval){ [ 965.151130] env[62692]: value = "task-1141515" [ 965.151130] env[62692]: _type = "Task" [ 965.151130] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.156415] env[62692]: DEBUG nova.network.neutron [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.165741] env[62692]: DEBUG oslo_vmware.api [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141515, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.181961] env[62692]: DEBUG oslo_vmware.api [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52d0a4de-b10c-975e-a51f-34a46ddf8f96, 'name': SearchDatastore_Task, 'duration_secs': 0.012133} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.182298] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.182541] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 965.182784] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.182934] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.183694] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 965.184502] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-296ead3e-49ca-4ee3-b0a2-e08a8805e854 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.195820] env[62692]: DEBUG nova.compute.manager [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 965.199457] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 965.200210] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 965.203981] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c15f08f4-0847-40d0-b517-dc02006702ad {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.213651] env[62692]: DEBUG oslo_vmware.api [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 965.213651] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5251b1ab-b54e-4f14-1cf4-e0d0287bc2d3" [ 965.213651] env[62692]: _type = "Task" [ 965.213651] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.226190] env[62692]: DEBUG oslo_vmware.api [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5251b1ab-b54e-4f14-1cf4-e0d0287bc2d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.251531] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7d2e151-f92d-4519-aeb3-a028ea1bcbb5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.264408] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7974b05-2830-4209-a593-5e591d916c7f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.304845] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2888dd30-0680-421c-975d-fa7b889f19b0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.316038] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df8e4402-fcdc-4fd4-a146-3604f69f47b3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.334652] env[62692]: DEBUG nova.compute.provider_tree [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.352695] env[62692]: DEBUG nova.network.neutron [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Successfully created port: daea3ee8-3a77-43be-b648-99133117b287 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 965.370507] env[62692]: DEBUG oslo_concurrency.lockutils [None req-fc481d96-844d-476f-93ea-68033636a9f6 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "43d5f0b7-c953-4ca8-ae9b-e45d326e008f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 1.527s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.399618] env[62692]: INFO nova.compute.manager [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Took 40.99 seconds to build instance. [ 965.430097] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141512, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.527489] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141513, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.596419] env[62692]: DEBUG oslo_vmware.api [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141514, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.659902] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Releasing lock "refresh_cache-abee88ff-5cf5-4bf1-91e0-93b19cf30046" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.659902] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Updated the network info_cache for instance {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 965.659902] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 965.660275] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 965.660469] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 965.664401] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 965.664760] env[62692]: DEBUG oslo_vmware.api [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141515, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.664965] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 965.665185] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 965.665327] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62692) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 965.665478] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager.update_available_resource {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 965.729933] env[62692]: DEBUG oslo_vmware.api [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5251b1ab-b54e-4f14-1cf4-e0d0287bc2d3, 'name': SearchDatastore_Task, 'duration_secs': 0.040066} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.731196] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e3d9993-1eae-4273-8c01-a23251ac2840 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.741100] env[62692]: DEBUG oslo_vmware.api [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 965.741100] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52626609-7a92-620a-2a98-1f7ea15cf901" [ 965.741100] env[62692]: _type = "Task" [ 965.741100] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.752596] env[62692]: DEBUG oslo_vmware.api [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52626609-7a92-620a-2a98-1f7ea15cf901, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.839154] env[62692]: DEBUG nova.scheduler.client.report [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 965.903159] env[62692]: DEBUG oslo_concurrency.lockutils [None req-99e0a85b-5012-4364-9463-9f42b9990cd4 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "668fb0ea-7eed-4198-943c-8f916ec11368" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.147s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.932745] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141512, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.726481} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.933038] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 46bce865-188b-487c-a73f-cf524059c797/46bce865-188b-487c-a73f-cf524059c797.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 965.933268] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 965.934274] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b3882a9d-4178-4d01-8e44-bcff235da573 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.944816] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 965.944816] env[62692]: value = "task-1141516" [ 965.944816] env[62692]: _type = "Task" [ 965.944816] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.954271] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141516, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.966885] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "250b9cc2-706e-4fbf-887c-1c80fc27a0ed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.966885] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "250b9cc2-706e-4fbf-887c-1c80fc27a0ed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.021692] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141513, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.095473] env[62692]: DEBUG oslo_vmware.api [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141514, 'name': PowerOnVM_Task, 'duration_secs': 0.589217} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.096897] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 966.099110] env[62692]: DEBUG nova.compute.manager [None req-06eacb14-2b37-4f2e-8d0a-75d97232e6e0 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 966.100329] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f866123c-e2b9-4adf-8d6a-0d61a7af6d26 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.164445] env[62692]: DEBUG oslo_vmware.api [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141515, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.738095} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.164716] env[62692]: INFO nova.virt.vmwareapi.ds_util [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 084ce29c-302a-45cc-b36f-0143eccee298/81eb64fb-5a0c-47b7-9948-7dcb01b911bd-rescue.vmdk. [ 966.165501] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1638b7f2-2368-4f0b-a5ee-93b69cdc8ba6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.169147] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.194276] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] 084ce29c-302a-45cc-b36f-0143eccee298/81eb64fb-5a0c-47b7-9948-7dcb01b911bd-rescue.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 966.194597] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da085738-0542-40e0-9093-32ae78ba93f4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.211219] env[62692]: DEBUG nova.compute.manager [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 966.222476] env[62692]: DEBUG oslo_vmware.api [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Waiting for the task: (returnval){ [ 966.222476] env[62692]: value = "task-1141517" [ 966.222476] env[62692]: _type = "Task" [ 966.222476] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.235235] env[62692]: DEBUG oslo_vmware.api [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141517, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.238564] env[62692]: DEBUG nova.virt.hardware [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 966.238782] env[62692]: DEBUG nova.virt.hardware [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 966.238998] env[62692]: DEBUG nova.virt.hardware [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 966.239289] env[62692]: DEBUG nova.virt.hardware [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 966.239502] env[62692]: DEBUG nova.virt.hardware [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 966.239652] env[62692]: DEBUG nova.virt.hardware [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 966.239888] env[62692]: DEBUG nova.virt.hardware [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 966.240070] env[62692]: DEBUG nova.virt.hardware [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 966.240256] env[62692]: DEBUG nova.virt.hardware [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 966.240800] env[62692]: DEBUG nova.virt.hardware [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 966.240800] env[62692]: DEBUG nova.virt.hardware [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 966.241489] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eae59ba1-f014-44f4-807c-7ca1bc1dfad6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.253790] env[62692]: DEBUG oslo_vmware.api [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52626609-7a92-620a-2a98-1f7ea15cf901, 'name': SearchDatastore_Task, 'duration_secs': 0.021064} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.256197] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.256487] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] f3e8be1c-d952-49b7-86a8-d0a6c56ee096/f3e8be1c-d952-49b7-86a8-d0a6c56ee096.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 966.256792] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aeeabe75-79f4-4da8-8e4e-dbaffd5da223 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.260804] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5748d06d-6e03-4c26-b74c-a9dbb42cb6c5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.277386] env[62692]: DEBUG oslo_vmware.api [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 966.277386] env[62692]: value = "task-1141518" [ 966.277386] env[62692]: _type = "Task" [ 966.277386] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.285948] env[62692]: DEBUG oslo_vmware.api [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141518, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.346213] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.160s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.348714] env[62692]: DEBUG oslo_concurrency.lockutils [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.375s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.350455] env[62692]: INFO nova.compute.claims [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 966.375914] env[62692]: INFO nova.scheduler.client.report [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Deleted allocations for instance 44699df6-5262-4700-a096-8c0e450323d2 [ 966.456491] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141516, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075158} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.456491] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 966.456491] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9625a2be-acc9-4a6d-8874-74fd99be6723 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.472073] env[62692]: DEBUG nova.compute.manager [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 966.486074] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 46bce865-188b-487c-a73f-cf524059c797/46bce865-188b-487c-a73f-cf524059c797.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 966.487079] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53f87d70-d5c5-4016-8d6f-a8f5d0fee2d4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.514978] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 966.514978] env[62692]: value = "task-1141519" [ 966.514978] env[62692]: _type = "Task" [ 966.514978] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.523954] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141513, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.527449] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141519, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.736433] env[62692]: DEBUG oslo_vmware.api [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141517, 'name': ReconfigVM_Task, 'duration_secs': 0.320303} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.736433] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Reconfigured VM instance instance-0000004e to attach disk [datastore2] 084ce29c-302a-45cc-b36f-0143eccee298/81eb64fb-5a0c-47b7-9948-7dcb01b911bd-rescue.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 966.737118] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a6fd238-e317-4627-9ce5-910d89779fc2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.763942] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-77d28337-2f8b-4720-afa9-3b6820e216b4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.783435] env[62692]: DEBUG oslo_vmware.api [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Waiting for the task: (returnval){ [ 966.783435] env[62692]: value = "task-1141520" [ 966.783435] env[62692]: _type = "Task" [ 966.783435] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.801316] env[62692]: DEBUG oslo_vmware.api [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141518, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.804879] env[62692]: DEBUG oslo_vmware.api [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141520, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.884667] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a3941aba-c6f5-4085-a9de-0a1b299fa26e tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "44699df6-5262-4700-a096-8c0e450323d2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.740s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.020027] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.025997] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141513, 'name': PowerOnVM_Task, 'duration_secs': 1.924706} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.026414] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 967.026674] env[62692]: DEBUG nova.compute.manager [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 967.028198] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59c93bf9-be4e-4f4f-a982-873ff28c37ec {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.036137] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141519, 'name': ReconfigVM_Task, 'duration_secs': 0.326268} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.038200] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 46bce865-188b-487c-a73f-cf524059c797/46bce865-188b-487c-a73f-cf524059c797.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 967.047140] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c0106429-fdb7-4df4-9ef4-dff5ef32e6fe {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.062418] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 967.062418] env[62692]: value = "task-1141521" [ 967.062418] env[62692]: _type = "Task" [ 967.062418] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.075076] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141521, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.116288] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "7183ec4a-992d-48f1-8dda-7f499c2f4e1d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.116585] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "7183ec4a-992d-48f1-8dda-7f499c2f4e1d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.294172] env[62692]: DEBUG oslo_vmware.api [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141518, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.95666} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.297277] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] f3e8be1c-d952-49b7-86a8-d0a6c56ee096/f3e8be1c-d952-49b7-86a8-d0a6c56ee096.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 967.297515] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 967.297777] env[62692]: DEBUG oslo_vmware.api [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141520, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.298035] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1ef17fd8-6b31-4f91-b10a-a1a7288e698b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.305204] env[62692]: DEBUG oslo_vmware.api [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 967.305204] env[62692]: value = "task-1141522" [ 967.305204] env[62692]: _type = "Task" [ 967.305204] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.316506] env[62692]: DEBUG oslo_vmware.api [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141522, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.351212] env[62692]: DEBUG nova.compute.manager [req-6665c615-f7e1-4efd-8742-e75e800df563 req-05dc98ec-ebad-4460-9987-8d8584734505 service nova] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Received event network-vif-plugged-daea3ee8-3a77-43be-b648-99133117b287 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 967.351464] env[62692]: DEBUG oslo_concurrency.lockutils [req-6665c615-f7e1-4efd-8742-e75e800df563 req-05dc98ec-ebad-4460-9987-8d8584734505 service nova] Acquiring lock "f7664593-1b7e-495f-8d0e-9137d7f180cc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.351710] env[62692]: DEBUG oslo_concurrency.lockutils [req-6665c615-f7e1-4efd-8742-e75e800df563 req-05dc98ec-ebad-4460-9987-8d8584734505 service nova] Lock "f7664593-1b7e-495f-8d0e-9137d7f180cc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.351843] env[62692]: DEBUG oslo_concurrency.lockutils [req-6665c615-f7e1-4efd-8742-e75e800df563 req-05dc98ec-ebad-4460-9987-8d8584734505 service nova] Lock "f7664593-1b7e-495f-8d0e-9137d7f180cc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.352021] env[62692]: DEBUG nova.compute.manager [req-6665c615-f7e1-4efd-8742-e75e800df563 req-05dc98ec-ebad-4460-9987-8d8584734505 service nova] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] No waiting events found dispatching network-vif-plugged-daea3ee8-3a77-43be-b648-99133117b287 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 967.352655] env[62692]: WARNING nova.compute.manager [req-6665c615-f7e1-4efd-8742-e75e800df563 req-05dc98ec-ebad-4460-9987-8d8584734505 service nova] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Received unexpected event network-vif-plugged-daea3ee8-3a77-43be-b648-99133117b287 for instance with vm_state building and task_state spawning. [ 967.463799] env[62692]: DEBUG nova.network.neutron [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Successfully updated port: daea3ee8-3a77-43be-b648-99133117b287 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 967.550344] env[62692]: INFO nova.compute.manager [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] bringing vm to original state: 'stopped' [ 967.564152] env[62692]: INFO nova.compute.manager [None req-efc920c5-a348-46ed-bd34-6001435f9df7 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Unrescuing [ 967.564438] env[62692]: DEBUG oslo_concurrency.lockutils [None req-efc920c5-a348-46ed-bd34-6001435f9df7 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "refresh_cache-3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.564599] env[62692]: DEBUG oslo_concurrency.lockutils [None req-efc920c5-a348-46ed-bd34-6001435f9df7 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquired lock "refresh_cache-3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.564792] env[62692]: DEBUG nova.network.neutron [None req-efc920c5-a348-46ed-bd34-6001435f9df7 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 967.581458] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141521, 'name': Rename_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.621036] env[62692]: DEBUG nova.compute.manager [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 967.795862] env[62692]: DEBUG oslo_vmware.api [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141520, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.816855] env[62692]: DEBUG oslo_vmware.api [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141522, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07636} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.817291] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 967.818220] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-392a7350-47c4-406b-8732-6ac239746aee {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.844998] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] f3e8be1c-d952-49b7-86a8-d0a6c56ee096/f3e8be1c-d952-49b7-86a8-d0a6c56ee096.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 967.848299] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7b4eeabc-bd37-4d31-bd50-85604d89d0a4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.872514] env[62692]: DEBUG oslo_vmware.api [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 967.872514] env[62692]: value = "task-1141523" [ 967.872514] env[62692]: _type = "Task" [ 967.872514] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.877878] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f12d4996-6d11-4dd5-aac5-28b2eaab86d9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.888865] env[62692]: DEBUG oslo_vmware.api [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141523, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.895089] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9333917-059a-4c1b-8df4-50aacb41170e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.934644] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fe27027-cb18-4e56-83f0-4fce87c861a5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.944497] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2e6536d-b38a-4b46-9718-aa7741e5d78b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.960160] env[62692]: DEBUG nova.compute.provider_tree [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 967.966133] env[62692]: DEBUG oslo_concurrency.lockutils [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "refresh_cache-f7664593-1b7e-495f-8d0e-9137d7f180cc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.966266] env[62692]: DEBUG oslo_concurrency.lockutils [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquired lock "refresh_cache-f7664593-1b7e-495f-8d0e-9137d7f180cc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.966432] env[62692]: DEBUG nova.network.neutron [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 968.085922] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141521, 'name': Rename_Task, 'duration_secs': 0.893626} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.086488] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 968.086912] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8c5888e7-4896-48af-ac66-9785138fc459 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.094781] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 968.094781] env[62692]: value = "task-1141524" [ 968.094781] env[62692]: _type = "Task" [ 968.094781] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.104122] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141524, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.159675] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.296818] env[62692]: DEBUG oslo_vmware.api [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141520, 'name': ReconfigVM_Task, 'duration_secs': 1.189234} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.297375] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 968.297728] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-832c85dd-e41d-4a9f-9618-ac49a44250a5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.305679] env[62692]: DEBUG oslo_vmware.api [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Waiting for the task: (returnval){ [ 968.305679] env[62692]: value = "task-1141525" [ 968.305679] env[62692]: _type = "Task" [ 968.305679] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.315162] env[62692]: DEBUG oslo_vmware.api [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141525, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.385052] env[62692]: DEBUG oslo_vmware.api [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141523, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.409698] env[62692]: DEBUG nova.network.neutron [None req-efc920c5-a348-46ed-bd34-6001435f9df7 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Updating instance_info_cache with network_info: [{"id": "2c7422d2-dfef-4010-a99a-60c510c54a3c", "address": "fa:16:3e:71:16:ae", "network": {"id": "17e72c3b-603f-4305-96ac-89a9b4c9c7e9", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-123501993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7db5465cb5db465b92dfa570572b6e80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c7422d2-df", "ovs_interfaceid": "2c7422d2-dfef-4010-a99a-60c510c54a3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.467747] env[62692]: DEBUG nova.scheduler.client.report [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 968.475881] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "64a8329d-e0e9-4c2d-bd1f-844ee40eb980" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.476499] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "64a8329d-e0e9-4c2d-bd1f-844ee40eb980" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.476807] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "64a8329d-e0e9-4c2d-bd1f-844ee40eb980-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.477081] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "64a8329d-e0e9-4c2d-bd1f-844ee40eb980-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.477429] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "64a8329d-e0e9-4c2d-bd1f-844ee40eb980-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.481499] env[62692]: INFO nova.compute.manager [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Terminating instance [ 968.487294] env[62692]: DEBUG nova.compute.manager [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 968.487648] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 968.489191] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d622263-3148-4b08-9a3b-c331388bad3d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.498563] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 968.499926] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a9a412d8-7e1d-47a3-adf3-a3d3defd0712 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.509040] env[62692]: DEBUG oslo_vmware.api [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 968.509040] env[62692]: value = "task-1141526" [ 968.509040] env[62692]: _type = "Task" [ 968.509040] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.523289] env[62692]: DEBUG oslo_vmware.api [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141526, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.543926] env[62692]: DEBUG nova.network.neutron [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 968.562925] env[62692]: DEBUG oslo_concurrency.lockutils [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "a668a569-a90e-4e59-8d2e-b0225745b500" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.563388] env[62692]: DEBUG oslo_concurrency.lockutils [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "a668a569-a90e-4e59-8d2e-b0225745b500" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.563835] env[62692]: DEBUG nova.compute.manager [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 968.564998] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b894d1b-65db-43bc-9046-7d69db2bac92 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.578607] env[62692]: DEBUG nova.compute.manager [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62692) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 968.581473] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 968.581966] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c22d2e44-b458-456f-a0a2-edf3809f18f4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.590933] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 968.590933] env[62692]: value = "task-1141527" [ 968.590933] env[62692]: _type = "Task" [ 968.590933] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.602857] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141527, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.609070] env[62692]: DEBUG oslo_vmware.api [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141524, 'name': PowerOnVM_Task, 'duration_secs': 0.510668} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.609483] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 968.610136] env[62692]: DEBUG nova.compute.manager [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 968.610763] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84c47f10-bc87-4e9e-85d4-324e259b3eff {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.739348] env[62692]: DEBUG nova.network.neutron [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Updating instance_info_cache with network_info: [{"id": "daea3ee8-3a77-43be-b648-99133117b287", "address": "fa:16:3e:58:96:36", "network": {"id": "d0e00428-1d04-4b63-8975-7f3085784358", "bridge": "br-int", "label": "tempest-ServersTestJSON-408035146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e13f8888d5e84eed8247101558e29a56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdaea3ee8-3a", "ovs_interfaceid": "daea3ee8-3a77-43be-b648-99133117b287", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.819527] env[62692]: DEBUG oslo_vmware.api [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141525, 'name': PowerOnVM_Task, 'duration_secs': 0.497536} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.819950] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 968.823123] env[62692]: DEBUG nova.compute.manager [None req-a76c68b8-4b7d-4b2a-bfec-88d851721e80 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 968.824018] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e485b40-1049-4eda-8725-4b063dde9dfa {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.888100] env[62692]: DEBUG oslo_vmware.api [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141523, 'name': ReconfigVM_Task, 'duration_secs': 0.67649} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.888390] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Reconfigured VM instance instance-00000050 to attach disk [datastore1] f3e8be1c-d952-49b7-86a8-d0a6c56ee096/f3e8be1c-d952-49b7-86a8-d0a6c56ee096.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 968.892806] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bca41ffa-034b-47c8-94f7-a14dea878dc8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.901887] env[62692]: DEBUG oslo_vmware.api [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 968.901887] env[62692]: value = "task-1141528" [ 968.901887] env[62692]: _type = "Task" [ 968.901887] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.911755] env[62692]: DEBUG oslo_vmware.api [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141528, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.912353] env[62692]: DEBUG oslo_concurrency.lockutils [None req-efc920c5-a348-46ed-bd34-6001435f9df7 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Releasing lock "refresh_cache-3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.912868] env[62692]: DEBUG nova.objects.instance [None req-efc920c5-a348-46ed-bd34-6001435f9df7 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lazy-loading 'flavor' on Instance uuid 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 968.984947] env[62692]: DEBUG oslo_concurrency.lockutils [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.636s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.985663] env[62692]: DEBUG nova.compute.manager [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 968.989154] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.814s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.989154] env[62692]: DEBUG nova.objects.instance [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Lazy-loading 'resources' on Instance uuid fac426b6-747a-455e-85a1-44e71b432180 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 969.020335] env[62692]: DEBUG oslo_vmware.api [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141526, 'name': PowerOffVM_Task, 'duration_secs': 0.329926} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.020335] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 969.020547] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 969.020780] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b3f05322-8832-4952-aef2-9fb2074010a9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.101483] env[62692]: DEBUG oslo_vmware.api [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141527, 'name': PowerOffVM_Task, 'duration_secs': 0.280254} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.102220] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 969.102520] env[62692]: DEBUG nova.compute.manager [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 969.103395] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be9a1a5a-af32-4cfd-bb72-f9af126a3688 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.134023] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 969.134023] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 969.134023] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Deleting the datastore file [datastore2] 64a8329d-e0e9-4c2d-bd1f-844ee40eb980 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 969.134601] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ca5c698c-3b64-456a-9867-7b35b0c4074b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.137222] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.143140] env[62692]: DEBUG oslo_vmware.api [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 969.143140] env[62692]: value = "task-1141530" [ 969.143140] env[62692]: _type = "Task" [ 969.143140] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.153212] env[62692]: DEBUG oslo_vmware.api [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141530, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.242184] env[62692]: DEBUG oslo_concurrency.lockutils [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Releasing lock "refresh_cache-f7664593-1b7e-495f-8d0e-9137d7f180cc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.242540] env[62692]: DEBUG nova.compute.manager [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Instance network_info: |[{"id": "daea3ee8-3a77-43be-b648-99133117b287", "address": "fa:16:3e:58:96:36", "network": {"id": "d0e00428-1d04-4b63-8975-7f3085784358", "bridge": "br-int", "label": "tempest-ServersTestJSON-408035146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e13f8888d5e84eed8247101558e29a56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdaea3ee8-3a", "ovs_interfaceid": "daea3ee8-3a77-43be-b648-99133117b287", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 969.243017] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:58:96:36', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'daea3ee8-3a77-43be-b648-99133117b287', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 969.250611] env[62692]: DEBUG oslo.service.loopingcall [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 969.250847] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 969.251093] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3787a927-bd51-448a-9903-495d61a4b716 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.275094] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 969.275094] env[62692]: value = "task-1141531" [ 969.275094] env[62692]: _type = "Task" [ 969.275094] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.284466] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141531, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.300284] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "f6085887-a5d3-4096-be38-d843bb54304a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.300522] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "f6085887-a5d3-4096-be38-d843bb54304a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.381614] env[62692]: DEBUG nova.compute.manager [req-25f74075-7406-4b99-ad4e-24c8d0c0efd7 req-e00b5e08-cf07-400f-bf7b-fd33fe9ce5b4 service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Received event network-changed-2deb4259-2f64-4f7c-9289-eba76375335a {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 969.381724] env[62692]: DEBUG nova.compute.manager [req-25f74075-7406-4b99-ad4e-24c8d0c0efd7 req-e00b5e08-cf07-400f-bf7b-fd33fe9ce5b4 service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Refreshing instance network info cache due to event network-changed-2deb4259-2f64-4f7c-9289-eba76375335a. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 969.381941] env[62692]: DEBUG oslo_concurrency.lockutils [req-25f74075-7406-4b99-ad4e-24c8d0c0efd7 req-e00b5e08-cf07-400f-bf7b-fd33fe9ce5b4 service nova] Acquiring lock "refresh_cache-668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.382097] env[62692]: DEBUG oslo_concurrency.lockutils [req-25f74075-7406-4b99-ad4e-24c8d0c0efd7 req-e00b5e08-cf07-400f-bf7b-fd33fe9ce5b4 service nova] Acquired lock "refresh_cache-668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.382269] env[62692]: DEBUG nova.network.neutron [req-25f74075-7406-4b99-ad4e-24c8d0c0efd7 req-e00b5e08-cf07-400f-bf7b-fd33fe9ce5b4 service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Refreshing network info cache for port 2deb4259-2f64-4f7c-9289-eba76375335a {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 969.413059] env[62692]: DEBUG oslo_vmware.api [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141528, 'name': Rename_Task, 'duration_secs': 0.413273} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.413059] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 969.413232] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-89674290-d3f7-4d18-9aaf-3216c653ecad {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.418106] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-152baad1-d2c4-44a5-941e-5c9af65ae7f1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.422571] env[62692]: DEBUG oslo_vmware.api [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 969.422571] env[62692]: value = "task-1141532" [ 969.422571] env[62692]: _type = "Task" [ 969.422571] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.444603] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-efc920c5-a348-46ed-bd34-6001435f9df7 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 969.445451] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f2ffeb2f-11a7-4848-8f72-1b0c4830e706 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.451015] env[62692]: DEBUG oslo_vmware.api [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141532, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.452405] env[62692]: DEBUG oslo_vmware.api [None req-efc920c5-a348-46ed-bd34-6001435f9df7 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 969.452405] env[62692]: value = "task-1141533" [ 969.452405] env[62692]: _type = "Task" [ 969.452405] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.461174] env[62692]: DEBUG oslo_vmware.api [None req-efc920c5-a348-46ed-bd34-6001435f9df7 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141533, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.493420] env[62692]: DEBUG nova.compute.utils [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 969.498441] env[62692]: DEBUG nova.compute.manager [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 969.498635] env[62692]: DEBUG nova.network.neutron [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 969.582917] env[62692]: DEBUG nova.policy [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c4e9433f9bac46c1b92f3f7ffe56fb6b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b4d2edc0134249a590f6a25c4d372315', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 969.620545] env[62692]: DEBUG oslo_concurrency.lockutils [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "a668a569-a90e-4e59-8d2e-b0225745b500" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.057s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.663989] env[62692]: DEBUG oslo_vmware.api [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141530, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.49395} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.664469] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 969.664687] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 969.664884] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 969.665308] env[62692]: INFO nova.compute.manager [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Took 1.18 seconds to destroy the instance on the hypervisor. [ 969.665411] env[62692]: DEBUG oslo.service.loopingcall [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 969.668231] env[62692]: DEBUG nova.compute.manager [-] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 969.668231] env[62692]: DEBUG nova.network.neutron [-] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 969.789427] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141531, 'name': CreateVM_Task, 'duration_secs': 0.489801} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.789427] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 969.789427] env[62692]: DEBUG oslo_concurrency.lockutils [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.789427] env[62692]: DEBUG oslo_concurrency.lockutils [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.789427] env[62692]: DEBUG oslo_concurrency.lockutils [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 969.789427] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-894a0d8b-3abf-4212-9b51-ec503ee98bbf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.798729] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 969.798729] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5278eaf5-7d69-14b9-3484-db08c6348b1d" [ 969.798729] env[62692]: _type = "Task" [ 969.798729] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.810909] env[62692]: DEBUG nova.compute.manager [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 969.814473] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5278eaf5-7d69-14b9-3484-db08c6348b1d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.934219] env[62692]: DEBUG oslo_vmware.api [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141532, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.963720] env[62692]: DEBUG oslo_vmware.api [None req-efc920c5-a348-46ed-bd34-6001435f9df7 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141533, 'name': PowerOffVM_Task, 'duration_secs': 0.325259} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.967289] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-efc920c5-a348-46ed-bd34-6001435f9df7 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 969.972230] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-efc920c5-a348-46ed-bd34-6001435f9df7 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Reconfiguring VM instance instance-00000040 to detach disk 2002 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 969.973689] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-15aa16a0-46bf-429c-859e-00ac34c629f0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.997333] env[62692]: DEBUG oslo_vmware.api [None req-efc920c5-a348-46ed-bd34-6001435f9df7 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 969.997333] env[62692]: value = "task-1141534" [ 969.997333] env[62692]: _type = "Task" [ 969.997333] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.004110] env[62692]: DEBUG nova.compute.manager [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 970.014838] env[62692]: DEBUG oslo_vmware.api [None req-efc920c5-a348-46ed-bd34-6001435f9df7 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141534, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.022543] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02c5caef-20c0-4037-affa-d75208735b91 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.029732] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3aa50ac-2410-4a43-88dd-4fd24349a636 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.071218] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aea17c65-9f16-488a-9650-5c8c22da5a4e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.089042] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24fc8213-ab0c-47f6-9211-d09bd06cc194 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.103844] env[62692]: DEBUG nova.compute.provider_tree [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 970.136671] env[62692]: DEBUG oslo_concurrency.lockutils [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.255296] env[62692]: DEBUG nova.network.neutron [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Successfully created port: 8842838a-68c4-4c35-b3dc-ba680c8f3732 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 970.277078] env[62692]: DEBUG nova.network.neutron [req-25f74075-7406-4b99-ad4e-24c8d0c0efd7 req-e00b5e08-cf07-400f-bf7b-fd33fe9ce5b4 service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Updated VIF entry in instance network info cache for port 2deb4259-2f64-4f7c-9289-eba76375335a. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 970.277643] env[62692]: DEBUG nova.network.neutron [req-25f74075-7406-4b99-ad4e-24c8d0c0efd7 req-e00b5e08-cf07-400f-bf7b-fd33fe9ce5b4 service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Updating instance_info_cache with network_info: [{"id": "2deb4259-2f64-4f7c-9289-eba76375335a", "address": "fa:16:3e:bd:3f:b3", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.135", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2deb4259-2f", "ovs_interfaceid": "2deb4259-2f64-4f7c-9289-eba76375335a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.310937] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5278eaf5-7d69-14b9-3484-db08c6348b1d, 'name': SearchDatastore_Task, 'duration_secs': 0.015565} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.311382] env[62692]: DEBUG oslo_concurrency.lockutils [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.311732] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 970.311994] env[62692]: DEBUG oslo_concurrency.lockutils [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.312165] env[62692]: DEBUG oslo_concurrency.lockutils [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.312430] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 970.312810] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9b3ccefa-11de-4201-a710-b2fb6550370a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.327236] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 970.327319] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 970.328092] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab8c8e64-7a98-4a71-a3b5-acba55bb0038 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.334980] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 970.334980] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]521e9849-cc5a-3ffc-4026-15bb3b830808" [ 970.334980] env[62692]: _type = "Task" [ 970.334980] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.336524] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.345996] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]521e9849-cc5a-3ffc-4026-15bb3b830808, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.434633] env[62692]: DEBUG oslo_vmware.api [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141532, 'name': PowerOnVM_Task, 'duration_secs': 0.608751} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.434956] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 970.435148] env[62692]: INFO nova.compute.manager [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Took 10.08 seconds to spawn the instance on the hypervisor. [ 970.435365] env[62692]: DEBUG nova.compute.manager [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 970.436243] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82783c03-c350-449a-8c2c-467c19499185 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.516026] env[62692]: DEBUG oslo_vmware.api [None req-efc920c5-a348-46ed-bd34-6001435f9df7 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141534, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.607504] env[62692]: DEBUG nova.scheduler.client.report [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 970.645063] env[62692]: DEBUG nova.compute.manager [req-3484e22d-84e6-477f-badb-a183594ab62c req-3f271d05-a00f-4f6f-ac54-c116563455db service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Received event network-changed-3fa9bd6e-092d-4e9f-a4ab-5113affb920b {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 970.645519] env[62692]: DEBUG nova.compute.manager [req-3484e22d-84e6-477f-badb-a183594ab62c req-3f271d05-a00f-4f6f-ac54-c116563455db service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Refreshing instance network info cache due to event network-changed-3fa9bd6e-092d-4e9f-a4ab-5113affb920b. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 970.646212] env[62692]: DEBUG oslo_concurrency.lockutils [req-3484e22d-84e6-477f-badb-a183594ab62c req-3f271d05-a00f-4f6f-ac54-c116563455db service nova] Acquiring lock "refresh_cache-084ce29c-302a-45cc-b36f-0143eccee298" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.646212] env[62692]: DEBUG oslo_concurrency.lockutils [req-3484e22d-84e6-477f-badb-a183594ab62c req-3f271d05-a00f-4f6f-ac54-c116563455db service nova] Acquired lock "refresh_cache-084ce29c-302a-45cc-b36f-0143eccee298" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.646613] env[62692]: DEBUG nova.network.neutron [req-3484e22d-84e6-477f-badb-a183594ab62c req-3f271d05-a00f-4f6f-ac54-c116563455db service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Refreshing network info cache for port 3fa9bd6e-092d-4e9f-a4ab-5113affb920b {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 970.654876] env[62692]: DEBUG nova.network.neutron [-] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.781422] env[62692]: DEBUG oslo_concurrency.lockutils [req-25f74075-7406-4b99-ad4e-24c8d0c0efd7 req-e00b5e08-cf07-400f-bf7b-fd33fe9ce5b4 service nova] Releasing lock "refresh_cache-668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.781741] env[62692]: DEBUG nova.compute.manager [req-25f74075-7406-4b99-ad4e-24c8d0c0efd7 req-e00b5e08-cf07-400f-bf7b-fd33fe9ce5b4 service nova] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Received event network-changed-daea3ee8-3a77-43be-b648-99133117b287 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 970.781924] env[62692]: DEBUG nova.compute.manager [req-25f74075-7406-4b99-ad4e-24c8d0c0efd7 req-e00b5e08-cf07-400f-bf7b-fd33fe9ce5b4 service nova] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Refreshing instance network info cache due to event network-changed-daea3ee8-3a77-43be-b648-99133117b287. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 970.782158] env[62692]: DEBUG oslo_concurrency.lockutils [req-25f74075-7406-4b99-ad4e-24c8d0c0efd7 req-e00b5e08-cf07-400f-bf7b-fd33fe9ce5b4 service nova] Acquiring lock "refresh_cache-f7664593-1b7e-495f-8d0e-9137d7f180cc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.782310] env[62692]: DEBUG oslo_concurrency.lockutils [req-25f74075-7406-4b99-ad4e-24c8d0c0efd7 req-e00b5e08-cf07-400f-bf7b-fd33fe9ce5b4 service nova] Acquired lock "refresh_cache-f7664593-1b7e-495f-8d0e-9137d7f180cc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.782522] env[62692]: DEBUG nova.network.neutron [req-25f74075-7406-4b99-ad4e-24c8d0c0efd7 req-e00b5e08-cf07-400f-bf7b-fd33fe9ce5b4 service nova] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Refreshing network info cache for port daea3ee8-3a77-43be-b648-99133117b287 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 970.850231] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]521e9849-cc5a-3ffc-4026-15bb3b830808, 'name': SearchDatastore_Task, 'duration_secs': 0.018985} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.851088] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c50d562-d463-40f3-8b87-197bf487e762 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.857307] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 970.857307] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52d45667-1ba0-1a08-1856-69404073f3c4" [ 970.857307] env[62692]: _type = "Task" [ 970.857307] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.866766] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52d45667-1ba0-1a08-1856-69404073f3c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.958517] env[62692]: INFO nova.compute.manager [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Took 39.40 seconds to build instance. [ 971.009066] env[62692]: DEBUG oslo_vmware.api [None req-efc920c5-a348-46ed-bd34-6001435f9df7 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141534, 'name': ReconfigVM_Task, 'duration_secs': 0.87797} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.009066] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-efc920c5-a348-46ed-bd34-6001435f9df7 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Reconfigured VM instance instance-00000040 to detach disk 2002 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 971.009253] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-efc920c5-a348-46ed-bd34-6001435f9df7 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 971.009615] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0a8bfc80-ce42-4319-8412-73c0f4712707 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.013636] env[62692]: DEBUG nova.compute.manager [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 971.018712] env[62692]: DEBUG oslo_vmware.api [None req-efc920c5-a348-46ed-bd34-6001435f9df7 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 971.018712] env[62692]: value = "task-1141535" [ 971.018712] env[62692]: _type = "Task" [ 971.018712] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.029969] env[62692]: DEBUG oslo_vmware.api [None req-efc920c5-a348-46ed-bd34-6001435f9df7 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141535, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.044335] env[62692]: DEBUG nova.virt.hardware [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 971.044335] env[62692]: DEBUG nova.virt.hardware [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 971.044335] env[62692]: DEBUG nova.virt.hardware [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 971.044548] env[62692]: DEBUG nova.virt.hardware [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 971.045857] env[62692]: DEBUG nova.virt.hardware [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 971.045857] env[62692]: DEBUG nova.virt.hardware [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 971.045857] env[62692]: DEBUG nova.virt.hardware [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 971.045857] env[62692]: DEBUG nova.virt.hardware [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 971.045857] env[62692]: DEBUG nova.virt.hardware [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 971.045857] env[62692]: DEBUG nova.virt.hardware [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 971.045857] env[62692]: DEBUG nova.virt.hardware [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 971.047776] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d846e908-52be-4d6e-805f-562f7269a3c0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.058282] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9afe3b38-6066-463f-9ac6-9bbcec342cea {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.114276] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.125s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.116563] env[62692]: DEBUG oslo_concurrency.lockutils [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.320s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.116844] env[62692]: DEBUG nova.objects.instance [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Lazy-loading 'resources' on Instance uuid 8a56e6df-eea1-41f4-9360-4f06d2f516a3 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 971.146752] env[62692]: INFO nova.scheduler.client.report [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Deleted allocations for instance fac426b6-747a-455e-85a1-44e71b432180 [ 971.157890] env[62692]: INFO nova.compute.manager [-] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Took 1.49 seconds to deallocate network for instance. [ 971.370381] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52d45667-1ba0-1a08-1856-69404073f3c4, 'name': SearchDatastore_Task, 'duration_secs': 0.028076} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.370670] env[62692]: DEBUG oslo_concurrency.lockutils [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.371015] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] f7664593-1b7e-495f-8d0e-9137d7f180cc/f7664593-1b7e-495f-8d0e-9137d7f180cc.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 971.371417] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8e97b29c-6b3a-44d7-be20-dc8f6bcf8f12 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.380136] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 971.380136] env[62692]: value = "task-1141536" [ 971.380136] env[62692]: _type = "Task" [ 971.380136] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.392520] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141536, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.461080] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a12de9f7-85aa-43f9-88cd-e8409b96f557 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "f3e8be1c-d952-49b7-86a8-d0a6c56ee096" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.549s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.537029] env[62692]: DEBUG oslo_vmware.api [None req-efc920c5-a348-46ed-bd34-6001435f9df7 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141535, 'name': PowerOnVM_Task, 'duration_secs': 0.481189} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.537468] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-efc920c5-a348-46ed-bd34-6001435f9df7 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 971.537981] env[62692]: DEBUG nova.compute.manager [None req-efc920c5-a348-46ed-bd34-6001435f9df7 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 971.538977] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1621935-a8fd-4fd8-bc82-65406972d364 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.542470] env[62692]: DEBUG nova.network.neutron [req-3484e22d-84e6-477f-badb-a183594ab62c req-3f271d05-a00f-4f6f-ac54-c116563455db service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Updated VIF entry in instance network info cache for port 3fa9bd6e-092d-4e9f-a4ab-5113affb920b. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 971.542774] env[62692]: DEBUG nova.network.neutron [req-3484e22d-84e6-477f-badb-a183594ab62c req-3f271d05-a00f-4f6f-ac54-c116563455db service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Updating instance_info_cache with network_info: [{"id": "3fa9bd6e-092d-4e9f-a4ab-5113affb920b", "address": "fa:16:3e:a2:35:15", "network": {"id": "bbdf7bb9-6b54-4e19-91c6-cb0650e1121b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-2011930120-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "4b111e98b02a401384b23e363372ed52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5", "external-id": "nsx-vlan-transportzone-685", "segmentation_id": 685, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3fa9bd6e-09", "ovs_interfaceid": "3fa9bd6e-092d-4e9f-a4ab-5113affb920b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.660784] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0317c-2fc5-45e6-a94e-28e207d9e645 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Lock "fac426b6-747a-455e-85a1-44e71b432180" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.332s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.666840] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.782764] env[62692]: DEBUG nova.compute.manager [req-eea00dd5-7630-479d-a2ba-dfbeaadd247c req-28e90c61-b149-42e8-b23d-abe93e4c67bb service nova] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Received event network-vif-deleted-99d28285-5d04-45ee-a29f-40f81babef5f {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 971.803159] env[62692]: DEBUG nova.network.neutron [req-25f74075-7406-4b99-ad4e-24c8d0c0efd7 req-e00b5e08-cf07-400f-bf7b-fd33fe9ce5b4 service nova] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Updated VIF entry in instance network info cache for port daea3ee8-3a77-43be-b648-99133117b287. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 971.803557] env[62692]: DEBUG nova.network.neutron [req-25f74075-7406-4b99-ad4e-24c8d0c0efd7 req-e00b5e08-cf07-400f-bf7b-fd33fe9ce5b4 service nova] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Updating instance_info_cache with network_info: [{"id": "daea3ee8-3a77-43be-b648-99133117b287", "address": "fa:16:3e:58:96:36", "network": {"id": "d0e00428-1d04-4b63-8975-7f3085784358", "bridge": "br-int", "label": "tempest-ServersTestJSON-408035146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e13f8888d5e84eed8247101558e29a56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdaea3ee8-3a", "ovs_interfaceid": "daea3ee8-3a77-43be-b648-99133117b287", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.892603] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141536, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.905897] env[62692]: DEBUG oslo_concurrency.lockutils [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "46bce865-188b-487c-a73f-cf524059c797" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.906195] env[62692]: DEBUG oslo_concurrency.lockutils [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "46bce865-188b-487c-a73f-cf524059c797" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.906419] env[62692]: DEBUG oslo_concurrency.lockutils [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "46bce865-188b-487c-a73f-cf524059c797-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.906622] env[62692]: DEBUG oslo_concurrency.lockutils [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "46bce865-188b-487c-a73f-cf524059c797-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.906780] env[62692]: DEBUG oslo_concurrency.lockutils [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "46bce865-188b-487c-a73f-cf524059c797-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.908913] env[62692]: INFO nova.compute.manager [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Terminating instance [ 971.911534] env[62692]: DEBUG nova.compute.manager [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 971.911864] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 971.912953] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77497c7-5497-4d74-80fe-7aebd3997093 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.924829] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 971.925135] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1c4e2f41-4a6f-42b2-85d6-9b653bd53ac0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.935980] env[62692]: DEBUG oslo_vmware.api [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 971.935980] env[62692]: value = "task-1141537" [ 971.935980] env[62692]: _type = "Task" [ 971.935980] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.946269] env[62692]: DEBUG oslo_vmware.api [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141537, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.997282] env[62692]: DEBUG oslo_concurrency.lockutils [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "a668a569-a90e-4e59-8d2e-b0225745b500" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.997624] env[62692]: DEBUG oslo_concurrency.lockutils [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "a668a569-a90e-4e59-8d2e-b0225745b500" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.997933] env[62692]: DEBUG oslo_concurrency.lockutils [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "a668a569-a90e-4e59-8d2e-b0225745b500-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.998137] env[62692]: DEBUG oslo_concurrency.lockutils [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "a668a569-a90e-4e59-8d2e-b0225745b500-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.998354] env[62692]: DEBUG oslo_concurrency.lockutils [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "a668a569-a90e-4e59-8d2e-b0225745b500-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.004288] env[62692]: INFO nova.compute.manager [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Terminating instance [ 972.007650] env[62692]: DEBUG nova.compute.manager [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 972.007995] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 972.009367] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a63ffaf3-dc3d-41cf-8f08-c14326de54b8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.019238] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 972.019666] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1e48817e-6f87-456c-8d9c-6015e2acc70b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.046071] env[62692]: DEBUG oslo_concurrency.lockutils [req-3484e22d-84e6-477f-badb-a183594ab62c req-3f271d05-a00f-4f6f-ac54-c116563455db service nova] Releasing lock "refresh_cache-084ce29c-302a-45cc-b36f-0143eccee298" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.100701] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 972.100937] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 972.101232] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Deleting the datastore file [datastore1] a668a569-a90e-4e59-8d2e-b0225745b500 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 972.101413] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-74ddbb22-23c0-4e0d-8776-1b82ac780237 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.114290] env[62692]: DEBUG oslo_vmware.api [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 972.114290] env[62692]: value = "task-1141539" [ 972.114290] env[62692]: _type = "Task" [ 972.114290] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.123492] env[62692]: DEBUG oslo_vmware.api [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141539, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.143088] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8f813af-ca2d-475c-a2a1-7e912b87a141 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.151767] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ff34659-d546-409b-b4d6-ad01e4cd222e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.190018] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fc2f136-bab1-408a-8833-131288d74b40 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.199322] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63b899ae-960f-47bb-925c-5ce3e6918234 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.214764] env[62692]: DEBUG nova.compute.provider_tree [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 972.309502] env[62692]: DEBUG oslo_concurrency.lockutils [req-25f74075-7406-4b99-ad4e-24c8d0c0efd7 req-e00b5e08-cf07-400f-bf7b-fd33fe9ce5b4 service nova] Releasing lock "refresh_cache-f7664593-1b7e-495f-8d0e-9137d7f180cc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.392246] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141536, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.448997] env[62692]: DEBUG oslo_vmware.api [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141537, 'name': PowerOffVM_Task, 'duration_secs': 0.27113} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.448997] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 972.448997] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 972.448997] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-435799b4-44e4-4674-8dc4-bf17a01981b5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.540110] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 972.540218] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 972.540492] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Deleting the datastore file [datastore1] 46bce865-188b-487c-a73f-cf524059c797 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 972.540844] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cf254cb5-6e19-44de-9f53-33c377ae539a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.549397] env[62692]: DEBUG oslo_vmware.api [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 972.549397] env[62692]: value = "task-1141541" [ 972.549397] env[62692]: _type = "Task" [ 972.549397] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.558847] env[62692]: DEBUG oslo_vmware.api [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141541, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.609606] env[62692]: DEBUG nova.compute.manager [req-f29602e9-430d-472a-bf81-60eeffae36d8 req-b0a8029e-bf7b-4fcc-b35c-ee6d092c8b6b service nova] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Received event network-vif-plugged-8842838a-68c4-4c35-b3dc-ba680c8f3732 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 972.609831] env[62692]: DEBUG oslo_concurrency.lockutils [req-f29602e9-430d-472a-bf81-60eeffae36d8 req-b0a8029e-bf7b-4fcc-b35c-ee6d092c8b6b service nova] Acquiring lock "16790505-cdf6-4937-8839-fa685a5f413e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.610064] env[62692]: DEBUG oslo_concurrency.lockutils [req-f29602e9-430d-472a-bf81-60eeffae36d8 req-b0a8029e-bf7b-4fcc-b35c-ee6d092c8b6b service nova] Lock "16790505-cdf6-4937-8839-fa685a5f413e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.610245] env[62692]: DEBUG oslo_concurrency.lockutils [req-f29602e9-430d-472a-bf81-60eeffae36d8 req-b0a8029e-bf7b-4fcc-b35c-ee6d092c8b6b service nova] Lock "16790505-cdf6-4937-8839-fa685a5f413e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.610525] env[62692]: DEBUG nova.compute.manager [req-f29602e9-430d-472a-bf81-60eeffae36d8 req-b0a8029e-bf7b-4fcc-b35c-ee6d092c8b6b service nova] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] No waiting events found dispatching network-vif-plugged-8842838a-68c4-4c35-b3dc-ba680c8f3732 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 972.610882] env[62692]: WARNING nova.compute.manager [req-f29602e9-430d-472a-bf81-60eeffae36d8 req-b0a8029e-bf7b-4fcc-b35c-ee6d092c8b6b service nova] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Received unexpected event network-vif-plugged-8842838a-68c4-4c35-b3dc-ba680c8f3732 for instance with vm_state building and task_state spawning. [ 972.625311] env[62692]: DEBUG oslo_vmware.api [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141539, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.652079] env[62692]: DEBUG nova.network.neutron [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Successfully updated port: 8842838a-68c4-4c35-b3dc-ba680c8f3732 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 972.718090] env[62692]: DEBUG nova.scheduler.client.report [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 972.894591] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141536, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.128529} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.895226] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] f7664593-1b7e-495f-8d0e-9137d7f180cc/f7664593-1b7e-495f-8d0e-9137d7f180cc.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 972.895556] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 972.895905] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e89bd514-91cd-4c22-99e9-968a13e3cbe6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.904616] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 972.904616] env[62692]: value = "task-1141542" [ 972.904616] env[62692]: _type = "Task" [ 972.904616] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.907879] env[62692]: DEBUG nova.compute.manager [req-791df929-432f-4d5e-9ec2-b5a812591daa req-2e678d02-207d-4d01-b513-b71cb32513d1 service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Received event network-changed-3fa9bd6e-092d-4e9f-a4ab-5113affb920b {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 972.908097] env[62692]: DEBUG nova.compute.manager [req-791df929-432f-4d5e-9ec2-b5a812591daa req-2e678d02-207d-4d01-b513-b71cb32513d1 service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Refreshing instance network info cache due to event network-changed-3fa9bd6e-092d-4e9f-a4ab-5113affb920b. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 972.908468] env[62692]: DEBUG oslo_concurrency.lockutils [req-791df929-432f-4d5e-9ec2-b5a812591daa req-2e678d02-207d-4d01-b513-b71cb32513d1 service nova] Acquiring lock "refresh_cache-084ce29c-302a-45cc-b36f-0143eccee298" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.908468] env[62692]: DEBUG oslo_concurrency.lockutils [req-791df929-432f-4d5e-9ec2-b5a812591daa req-2e678d02-207d-4d01-b513-b71cb32513d1 service nova] Acquired lock "refresh_cache-084ce29c-302a-45cc-b36f-0143eccee298" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.908693] env[62692]: DEBUG nova.network.neutron [req-791df929-432f-4d5e-9ec2-b5a812591daa req-2e678d02-207d-4d01-b513-b71cb32513d1 service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Refreshing network info cache for port 3fa9bd6e-092d-4e9f-a4ab-5113affb920b {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 972.923052] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141542, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.060611] env[62692]: DEBUG oslo_vmware.api [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141541, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.346513} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.060892] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 973.061103] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 973.061296] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 973.061479] env[62692]: INFO nova.compute.manager [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Took 1.15 seconds to destroy the instance on the hypervisor. [ 973.061808] env[62692]: DEBUG oslo.service.loopingcall [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 973.062089] env[62692]: DEBUG nova.compute.manager [-] [instance: 46bce865-188b-487c-a73f-cf524059c797] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 973.062607] env[62692]: DEBUG nova.network.neutron [-] [instance: 46bce865-188b-487c-a73f-cf524059c797] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 973.128023] env[62692]: DEBUG oslo_vmware.api [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141539, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.551209} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.128023] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 973.128023] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 973.128023] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 973.128023] env[62692]: INFO nova.compute.manager [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Took 1.12 seconds to destroy the instance on the hypervisor. [ 973.128023] env[62692]: DEBUG oslo.service.loopingcall [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 973.128023] env[62692]: DEBUG nova.compute.manager [-] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 973.128023] env[62692]: DEBUG nova.network.neutron [-] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 973.157447] env[62692]: DEBUG oslo_concurrency.lockutils [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Acquiring lock "refresh_cache-16790505-cdf6-4937-8839-fa685a5f413e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.157564] env[62692]: DEBUG oslo_concurrency.lockutils [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Acquired lock "refresh_cache-16790505-cdf6-4937-8839-fa685a5f413e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.157653] env[62692]: DEBUG nova.network.neutron [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 973.223791] env[62692]: DEBUG oslo_concurrency.lockutils [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.107s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.227482] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.078s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.231019] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.231019] env[62692]: DEBUG oslo_concurrency.lockutils [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.954s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.231019] env[62692]: DEBUG nova.objects.instance [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lazy-loading 'resources' on Instance uuid abee88ff-5cf5-4bf1-91e0-93b19cf30046 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.250483] env[62692]: INFO nova.scheduler.client.report [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Deleted allocations for instance 2ccb76cf-f641-4306-a137-fb2417285df9 [ 973.254521] env[62692]: INFO nova.scheduler.client.report [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Deleted allocations for instance 8a56e6df-eea1-41f4-9360-4f06d2f516a3 [ 973.379814] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquiring lock "fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.380159] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Lock "fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.380389] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquiring lock "fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.380549] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Lock "fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.380729] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Lock "fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.383884] env[62692]: INFO nova.compute.manager [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Terminating instance [ 973.385691] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquiring lock "refresh_cache-fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.385772] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquired lock "refresh_cache-fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.385943] env[62692]: DEBUG nova.network.neutron [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 973.421375] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141542, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07446} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.422025] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 973.423899] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-569afc40-0bd8-4b5d-8c3e-ecb41f2901d3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.451693] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] f7664593-1b7e-495f-8d0e-9137d7f180cc/f7664593-1b7e-495f-8d0e-9137d7f180cc.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 973.452455] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b33d0470-f18d-45e9-8075-c2cf55574a10 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.473047] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 973.473047] env[62692]: value = "task-1141543" [ 973.473047] env[62692]: _type = "Task" [ 973.473047] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.484322] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141543, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.714730] env[62692]: DEBUG nova.network.neutron [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 973.766236] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3495813d-4944-407d-aa5a-19ba1605f9a9 tempest-MigrationsAdminTest-1774112206 tempest-MigrationsAdminTest-1774112206-project-member] Lock "2ccb76cf-f641-4306-a137-fb2417285df9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.002s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.767343] env[62692]: DEBUG oslo_concurrency.lockutils [None req-fd16bd26-ea0d-4894-9267-2a8216cbb244 tempest-ServerMetadataNegativeTestJSON-343793513 tempest-ServerMetadataNegativeTestJSON-343793513-project-member] Lock "8a56e6df-eea1-41f4-9360-4f06d2f516a3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.116s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.924379] env[62692]: DEBUG nova.network.neutron [req-791df929-432f-4d5e-9ec2-b5a812591daa req-2e678d02-207d-4d01-b513-b71cb32513d1 service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Updated VIF entry in instance network info cache for port 3fa9bd6e-092d-4e9f-a4ab-5113affb920b. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 973.924760] env[62692]: DEBUG nova.network.neutron [req-791df929-432f-4d5e-9ec2-b5a812591daa req-2e678d02-207d-4d01-b513-b71cb32513d1 service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Updating instance_info_cache with network_info: [{"id": "3fa9bd6e-092d-4e9f-a4ab-5113affb920b", "address": "fa:16:3e:a2:35:15", "network": {"id": "bbdf7bb9-6b54-4e19-91c6-cb0650e1121b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-2011930120-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "4b111e98b02a401384b23e363372ed52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5", "external-id": "nsx-vlan-transportzone-685", "segmentation_id": 685, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3fa9bd6e-09", "ovs_interfaceid": "3fa9bd6e-092d-4e9f-a4ab-5113affb920b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.929511] env[62692]: DEBUG nova.network.neutron [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 973.985950] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141543, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.016517] env[62692]: DEBUG nova.network.neutron [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Updating instance_info_cache with network_info: [{"id": "8842838a-68c4-4c35-b3dc-ba680c8f3732", "address": "fa:16:3e:93:81:b5", "network": {"id": "1a5b25dc-fc9f-4cd6-a14c-8227afea6670", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-568391236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4d2edc0134249a590f6a25c4d372315", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8fedd232-bfc1-4e7f-bd5e-c43ef8f2f08a", "external-id": "nsx-vlan-transportzone-925", "segmentation_id": 925, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8842838a-68", "ovs_interfaceid": "8842838a-68c4-4c35-b3dc-ba680c8f3732", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.037230] env[62692]: DEBUG nova.network.neutron [-] [instance: 46bce865-188b-487c-a73f-cf524059c797] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.049332] env[62692]: DEBUG nova.network.neutron [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.082512] env[62692]: DEBUG nova.network.neutron [-] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.241019] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dcf1612-3a19-422c-a639-fa7672f948d3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.253568] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aef94c8-ac25-401e-904d-1f8b0ead4d72 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.290536] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6d914d0-7349-4433-b6ad-8df47406fff3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.298949] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f2af522-f031-4249-bf47-4595e90e3d89 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.314336] env[62692]: DEBUG nova.compute.provider_tree [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 974.342155] env[62692]: DEBUG nova.compute.manager [req-8080e5c8-52a8-457e-880d-09b7013d6538 req-2cc2af96-4f6f-4f10-875a-10190453eee3 service nova] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Received event network-changed-1b3f2a08-aff4-4b7e-8b44-bc9033d93162 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 974.342665] env[62692]: DEBUG nova.compute.manager [req-8080e5c8-52a8-457e-880d-09b7013d6538 req-2cc2af96-4f6f-4f10-875a-10190453eee3 service nova] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Refreshing instance network info cache due to event network-changed-1b3f2a08-aff4-4b7e-8b44-bc9033d93162. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 974.343057] env[62692]: DEBUG oslo_concurrency.lockutils [req-8080e5c8-52a8-457e-880d-09b7013d6538 req-2cc2af96-4f6f-4f10-875a-10190453eee3 service nova] Acquiring lock "refresh_cache-f3e8be1c-d952-49b7-86a8-d0a6c56ee096" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.343336] env[62692]: DEBUG oslo_concurrency.lockutils [req-8080e5c8-52a8-457e-880d-09b7013d6538 req-2cc2af96-4f6f-4f10-875a-10190453eee3 service nova] Acquired lock "refresh_cache-f3e8be1c-d952-49b7-86a8-d0a6c56ee096" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.343605] env[62692]: DEBUG nova.network.neutron [req-8080e5c8-52a8-457e-880d-09b7013d6538 req-2cc2af96-4f6f-4f10-875a-10190453eee3 service nova] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Refreshing network info cache for port 1b3f2a08-aff4-4b7e-8b44-bc9033d93162 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 974.427422] env[62692]: DEBUG oslo_concurrency.lockutils [req-791df929-432f-4d5e-9ec2-b5a812591daa req-2e678d02-207d-4d01-b513-b71cb32513d1 service nova] Releasing lock "refresh_cache-084ce29c-302a-45cc-b36f-0143eccee298" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.486959] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141543, 'name': ReconfigVM_Task, 'duration_secs': 0.781225} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.487406] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Reconfigured VM instance instance-00000051 to attach disk [datastore1] f7664593-1b7e-495f-8d0e-9137d7f180cc/f7664593-1b7e-495f-8d0e-9137d7f180cc.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 974.488328] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-85057804-51d1-46a0-9b32-317d213075dd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.496024] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 974.496024] env[62692]: value = "task-1141544" [ 974.496024] env[62692]: _type = "Task" [ 974.496024] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.506231] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141544, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.521052] env[62692]: DEBUG oslo_concurrency.lockutils [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Releasing lock "refresh_cache-16790505-cdf6-4937-8839-fa685a5f413e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.521720] env[62692]: DEBUG nova.compute.manager [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Instance network_info: |[{"id": "8842838a-68c4-4c35-b3dc-ba680c8f3732", "address": "fa:16:3e:93:81:b5", "network": {"id": "1a5b25dc-fc9f-4cd6-a14c-8227afea6670", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-568391236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4d2edc0134249a590f6a25c4d372315", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8fedd232-bfc1-4e7f-bd5e-c43ef8f2f08a", "external-id": "nsx-vlan-transportzone-925", "segmentation_id": 925, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8842838a-68", "ovs_interfaceid": "8842838a-68c4-4c35-b3dc-ba680c8f3732", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 974.522255] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:93:81:b5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8fedd232-bfc1-4e7f-bd5e-c43ef8f2f08a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8842838a-68c4-4c35-b3dc-ba680c8f3732', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 974.530571] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Creating folder: Project (b4d2edc0134249a590f6a25c4d372315). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 974.531132] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fae059f9-4e44-48ff-987d-9182b8548b7d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.542105] env[62692]: INFO nova.compute.manager [-] [instance: 46bce865-188b-487c-a73f-cf524059c797] Took 1.48 seconds to deallocate network for instance. [ 974.547355] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Created folder: Project (b4d2edc0134249a590f6a25c4d372315) in parent group-v248868. [ 974.547642] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Creating folder: Instances. Parent ref: group-v249008. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 974.551276] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-10c3ad46-a955-4ae9-9a2b-a17a7c57d22e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.553613] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Releasing lock "refresh_cache-fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.554072] env[62692]: DEBUG nova.compute.manager [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 974.554382] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 974.555393] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13e1124d-a97f-4cbc-afff-2036b3f8a9cb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.566731] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 974.568320] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c1f43e12-3e4c-4c03-8496-296e967dc0e5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.570112] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Created folder: Instances in parent group-v249008. [ 974.570439] env[62692]: DEBUG oslo.service.loopingcall [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 974.570713] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 974.571342] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9fa66bae-0c52-4673-aa34-0a2c3154a71c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.589266] env[62692]: INFO nova.compute.manager [-] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Took 1.46 seconds to deallocate network for instance. [ 974.589921] env[62692]: DEBUG oslo_vmware.api [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 974.589921] env[62692]: value = "task-1141547" [ 974.589921] env[62692]: _type = "Task" [ 974.589921] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.600802] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 974.600802] env[62692]: value = "task-1141548" [ 974.600802] env[62692]: _type = "Task" [ 974.600802] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.610521] env[62692]: DEBUG oslo_vmware.api [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141547, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.618125] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141548, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.819286] env[62692]: DEBUG nova.scheduler.client.report [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 975.012905] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141544, 'name': Rename_Task, 'duration_secs': 0.236163} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.012905] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 975.013154] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-85f5cada-46aa-4933-bbfb-e0c6b4befdf2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.025017] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 975.025017] env[62692]: value = "task-1141549" [ 975.025017] env[62692]: _type = "Task" [ 975.025017] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.026718] env[62692]: DEBUG nova.compute.manager [req-3bbad77a-6675-43f1-b225-47b2aa01c27b req-41a679b3-f0f6-4c05-b181-0fa2792f6618 service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Received event network-changed-3fa9bd6e-092d-4e9f-a4ab-5113affb920b {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 975.026718] env[62692]: DEBUG nova.compute.manager [req-3bbad77a-6675-43f1-b225-47b2aa01c27b req-41a679b3-f0f6-4c05-b181-0fa2792f6618 service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Refreshing instance network info cache due to event network-changed-3fa9bd6e-092d-4e9f-a4ab-5113affb920b. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 975.026899] env[62692]: DEBUG oslo_concurrency.lockutils [req-3bbad77a-6675-43f1-b225-47b2aa01c27b req-41a679b3-f0f6-4c05-b181-0fa2792f6618 service nova] Acquiring lock "refresh_cache-084ce29c-302a-45cc-b36f-0143eccee298" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.027060] env[62692]: DEBUG oslo_concurrency.lockutils [req-3bbad77a-6675-43f1-b225-47b2aa01c27b req-41a679b3-f0f6-4c05-b181-0fa2792f6618 service nova] Acquired lock "refresh_cache-084ce29c-302a-45cc-b36f-0143eccee298" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.027228] env[62692]: DEBUG nova.network.neutron [req-3bbad77a-6675-43f1-b225-47b2aa01c27b req-41a679b3-f0f6-4c05-b181-0fa2792f6618 service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Refreshing network info cache for port 3fa9bd6e-092d-4e9f-a4ab-5113affb920b {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 975.038814] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141549, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.054424] env[62692]: DEBUG oslo_concurrency.lockutils [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.106512] env[62692]: DEBUG oslo_concurrency.lockutils [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.107450] env[62692]: DEBUG oslo_vmware.api [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141547, 'name': PowerOffVM_Task, 'duration_secs': 0.198316} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.111777] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 975.112124] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 975.112388] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cdad1a7a-ff03-4cf3-bde2-3ab7ecbd40bf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.124465] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141548, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.147663] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 975.147912] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 975.148116] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Deleting the datastore file [datastore2] fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 975.148402] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-825b1318-ad94-41ba-bbd0-35611f0557b6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.159627] env[62692]: DEBUG oslo_vmware.api [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for the task: (returnval){ [ 975.159627] env[62692]: value = "task-1141551" [ 975.159627] env[62692]: _type = "Task" [ 975.159627] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.170919] env[62692]: DEBUG oslo_vmware.api [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141551, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.278381] env[62692]: DEBUG nova.network.neutron [req-8080e5c8-52a8-457e-880d-09b7013d6538 req-2cc2af96-4f6f-4f10-875a-10190453eee3 service nova] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Updated VIF entry in instance network info cache for port 1b3f2a08-aff4-4b7e-8b44-bc9033d93162. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 975.279273] env[62692]: DEBUG nova.network.neutron [req-8080e5c8-52a8-457e-880d-09b7013d6538 req-2cc2af96-4f6f-4f10-875a-10190453eee3 service nova] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Updating instance_info_cache with network_info: [{"id": "1b3f2a08-aff4-4b7e-8b44-bc9033d93162", "address": "fa:16:3e:de:6b:59", "network": {"id": "e9c538a3-7723-42cb-85b7-5a1597e49cef", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1878627765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71b00c566cea4b21811ea1e1ef5ba0f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b3f2a08-af", "ovs_interfaceid": "1b3f2a08-aff4-4b7e-8b44-bc9033d93162", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.331137] env[62692]: DEBUG oslo_concurrency.lockutils [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.101s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.334127] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 27.918s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.334420] env[62692]: DEBUG nova.objects.instance [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62692) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 975.363994] env[62692]: INFO nova.scheduler.client.report [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Deleted allocations for instance abee88ff-5cf5-4bf1-91e0-93b19cf30046 [ 975.546113] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141549, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.619022] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141548, 'name': CreateVM_Task, 'duration_secs': 0.621597} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.619282] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 975.621448] env[62692]: DEBUG oslo_concurrency.lockutils [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.621448] env[62692]: DEBUG oslo_concurrency.lockutils [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.621448] env[62692]: DEBUG oslo_concurrency.lockutils [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 975.621448] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a357664d-bf3d-4893-be1d-f2075f087587 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.626667] env[62692]: DEBUG oslo_vmware.api [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Waiting for the task: (returnval){ [ 975.626667] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]524ec0cf-2833-af01-7a76-0601cb66dbd8" [ 975.626667] env[62692]: _type = "Task" [ 975.626667] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.638181] env[62692]: DEBUG oslo_vmware.api [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]524ec0cf-2833-af01-7a76-0601cb66dbd8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.669715] env[62692]: DEBUG oslo_vmware.api [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141551, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.782430] env[62692]: DEBUG oslo_concurrency.lockutils [req-8080e5c8-52a8-457e-880d-09b7013d6538 req-2cc2af96-4f6f-4f10-875a-10190453eee3 service nova] Releasing lock "refresh_cache-f3e8be1c-d952-49b7-86a8-d0a6c56ee096" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.785529] env[62692]: DEBUG nova.compute.manager [req-8080e5c8-52a8-457e-880d-09b7013d6538 req-2cc2af96-4f6f-4f10-875a-10190453eee3 service nova] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Received event network-changed-8842838a-68c4-4c35-b3dc-ba680c8f3732 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 975.785529] env[62692]: DEBUG nova.compute.manager [req-8080e5c8-52a8-457e-880d-09b7013d6538 req-2cc2af96-4f6f-4f10-875a-10190453eee3 service nova] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Refreshing instance network info cache due to event network-changed-8842838a-68c4-4c35-b3dc-ba680c8f3732. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 975.785529] env[62692]: DEBUG oslo_concurrency.lockutils [req-8080e5c8-52a8-457e-880d-09b7013d6538 req-2cc2af96-4f6f-4f10-875a-10190453eee3 service nova] Acquiring lock "refresh_cache-16790505-cdf6-4937-8839-fa685a5f413e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.785529] env[62692]: DEBUG oslo_concurrency.lockutils [req-8080e5c8-52a8-457e-880d-09b7013d6538 req-2cc2af96-4f6f-4f10-875a-10190453eee3 service nova] Acquired lock "refresh_cache-16790505-cdf6-4937-8839-fa685a5f413e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.785529] env[62692]: DEBUG nova.network.neutron [req-8080e5c8-52a8-457e-880d-09b7013d6538 req-2cc2af96-4f6f-4f10-875a-10190453eee3 service nova] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Refreshing network info cache for port 8842838a-68c4-4c35-b3dc-ba680c8f3732 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 975.795693] env[62692]: DEBUG nova.network.neutron [req-3bbad77a-6675-43f1-b225-47b2aa01c27b req-41a679b3-f0f6-4c05-b181-0fa2792f6618 service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Updated VIF entry in instance network info cache for port 3fa9bd6e-092d-4e9f-a4ab-5113affb920b. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 975.796089] env[62692]: DEBUG nova.network.neutron [req-3bbad77a-6675-43f1-b225-47b2aa01c27b req-41a679b3-f0f6-4c05-b181-0fa2792f6618 service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Updating instance_info_cache with network_info: [{"id": "3fa9bd6e-092d-4e9f-a4ab-5113affb920b", "address": "fa:16:3e:a2:35:15", "network": {"id": "bbdf7bb9-6b54-4e19-91c6-cb0650e1121b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-2011930120-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "4b111e98b02a401384b23e363372ed52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5", "external-id": "nsx-vlan-transportzone-685", "segmentation_id": 685, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3fa9bd6e-09", "ovs_interfaceid": "3fa9bd6e-092d-4e9f-a4ab-5113affb920b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.881513] env[62692]: DEBUG oslo_concurrency.lockutils [None req-14437cf4-6680-4b79-bc2e-0bb9101e9df8 tempest-VolumesAdminNegativeTest-1494727238 tempest-VolumesAdminNegativeTest-1494727238-project-member] Lock "abee88ff-5cf5-4bf1-91e0-93b19cf30046" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.309s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.037265] env[62692]: DEBUG oslo_vmware.api [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141549, 'name': PowerOnVM_Task, 'duration_secs': 0.821634} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.037635] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 976.037933] env[62692]: INFO nova.compute.manager [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Took 9.83 seconds to spawn the instance on the hypervisor. [ 976.038185] env[62692]: DEBUG nova.compute.manager [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 976.039690] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d595cd9c-d576-4be4-8952-e215676a20f9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.143210] env[62692]: DEBUG oslo_vmware.api [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]524ec0cf-2833-af01-7a76-0601cb66dbd8, 'name': SearchDatastore_Task, 'duration_secs': 0.012237} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.143210] env[62692]: DEBUG oslo_concurrency.lockutils [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.143210] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 976.143210] env[62692]: DEBUG oslo_concurrency.lockutils [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.143210] env[62692]: DEBUG oslo_concurrency.lockutils [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.143210] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 976.143210] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4e966782-47d5-4f32-883b-2124d8832825 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.152768] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 976.152959] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 976.153704] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2d7e0f6-df11-4ca8-a7fd-97c0f50a9971 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.161878] env[62692]: DEBUG oslo_vmware.api [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Waiting for the task: (returnval){ [ 976.161878] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]524170ac-745b-aded-2a42-299a23565183" [ 976.161878] env[62692]: _type = "Task" [ 976.161878] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.175027] env[62692]: DEBUG oslo_vmware.api [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Task: {'id': task-1141551, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.566747} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.178637] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 976.178847] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 976.179041] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 976.179229] env[62692]: INFO nova.compute.manager [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Took 1.62 seconds to destroy the instance on the hypervisor. [ 976.179544] env[62692]: DEBUG oslo.service.loopingcall [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 976.179770] env[62692]: DEBUG oslo_vmware.api [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]524170ac-745b-aded-2a42-299a23565183, 'name': SearchDatastore_Task, 'duration_secs': 0.009893} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.179976] env[62692]: DEBUG nova.compute.manager [-] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 976.180086] env[62692]: DEBUG nova.network.neutron [-] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 976.182333] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6043e22d-d3e6-4882-ac09-5259804c4388 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.189959] env[62692]: DEBUG oslo_vmware.api [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Waiting for the task: (returnval){ [ 976.189959] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52ecf910-61a7-f561-5c8b-97b01fe875c2" [ 976.189959] env[62692]: _type = "Task" [ 976.189959] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.199633] env[62692]: DEBUG oslo_vmware.api [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52ecf910-61a7-f561-5c8b-97b01fe875c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.200610] env[62692]: DEBUG nova.network.neutron [-] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 976.301067] env[62692]: DEBUG oslo_concurrency.lockutils [req-3bbad77a-6675-43f1-b225-47b2aa01c27b req-41a679b3-f0f6-4c05-b181-0fa2792f6618 service nova] Releasing lock "refresh_cache-084ce29c-302a-45cc-b36f-0143eccee298" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.301067] env[62692]: DEBUG nova.compute.manager [req-3bbad77a-6675-43f1-b225-47b2aa01c27b req-41a679b3-f0f6-4c05-b181-0fa2792f6618 service nova] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Received event network-changed-2c7422d2-dfef-4010-a99a-60c510c54a3c {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 976.301067] env[62692]: DEBUG nova.compute.manager [req-3bbad77a-6675-43f1-b225-47b2aa01c27b req-41a679b3-f0f6-4c05-b181-0fa2792f6618 service nova] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Refreshing instance network info cache due to event network-changed-2c7422d2-dfef-4010-a99a-60c510c54a3c. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 976.301417] env[62692]: DEBUG oslo_concurrency.lockutils [req-3bbad77a-6675-43f1-b225-47b2aa01c27b req-41a679b3-f0f6-4c05-b181-0fa2792f6618 service nova] Acquiring lock "refresh_cache-3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.301417] env[62692]: DEBUG oslo_concurrency.lockutils [req-3bbad77a-6675-43f1-b225-47b2aa01c27b req-41a679b3-f0f6-4c05-b181-0fa2792f6618 service nova] Acquired lock "refresh_cache-3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.301483] env[62692]: DEBUG nova.network.neutron [req-3bbad77a-6675-43f1-b225-47b2aa01c27b req-41a679b3-f0f6-4c05-b181-0fa2792f6618 service nova] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Refreshing network info cache for port 2c7422d2-dfef-4010-a99a-60c510c54a3c {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 976.338810] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Acquiring lock "084ce29c-302a-45cc-b36f-0143eccee298" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.339758] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Lock "084ce29c-302a-45cc-b36f-0143eccee298" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.340123] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Acquiring lock "084ce29c-302a-45cc-b36f-0143eccee298-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.340377] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Lock "084ce29c-302a-45cc-b36f-0143eccee298-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.340619] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Lock "084ce29c-302a-45cc-b36f-0143eccee298-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.342807] env[62692]: INFO nova.compute.manager [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Terminating instance [ 976.345078] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f364fd4b-062d-4585-91d6-e5e9ef3ae9f4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.352016] env[62692]: DEBUG nova.compute.manager [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 976.352016] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 976.352016] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.803s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.352016] env[62692]: DEBUG nova.objects.instance [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Lazy-loading 'resources' on Instance uuid e2b3b062-10e2-4a9c-b6b3-f0786de07b79 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 976.352016] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca047ed5-b2b9-4cd7-bc1d-709e5ba443bf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.363639] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 976.365595] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-74ae859b-017b-446e-88ba-afaca853b234 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.376083] env[62692]: DEBUG oslo_vmware.api [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Waiting for the task: (returnval){ [ 976.376083] env[62692]: value = "task-1141552" [ 976.376083] env[62692]: _type = "Task" [ 976.376083] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.390923] env[62692]: DEBUG oslo_vmware.api [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141552, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.468421] env[62692]: DEBUG nova.compute.manager [req-f840c916-94c7-454c-a318-8d9ef7704ce1 req-66aca2e2-7447-48d5-af4c-5202281c23d7 service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Received event network-changed-3fa9bd6e-092d-4e9f-a4ab-5113affb920b {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 976.468635] env[62692]: DEBUG nova.compute.manager [req-f840c916-94c7-454c-a318-8d9ef7704ce1 req-66aca2e2-7447-48d5-af4c-5202281c23d7 service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Refreshing instance network info cache due to event network-changed-3fa9bd6e-092d-4e9f-a4ab-5113affb920b. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 976.468891] env[62692]: DEBUG oslo_concurrency.lockutils [req-f840c916-94c7-454c-a318-8d9ef7704ce1 req-66aca2e2-7447-48d5-af4c-5202281c23d7 service nova] Acquiring lock "refresh_cache-084ce29c-302a-45cc-b36f-0143eccee298" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.469401] env[62692]: DEBUG oslo_concurrency.lockutils [req-f840c916-94c7-454c-a318-8d9ef7704ce1 req-66aca2e2-7447-48d5-af4c-5202281c23d7 service nova] Acquired lock "refresh_cache-084ce29c-302a-45cc-b36f-0143eccee298" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.469401] env[62692]: DEBUG nova.network.neutron [req-f840c916-94c7-454c-a318-8d9ef7704ce1 req-66aca2e2-7447-48d5-af4c-5202281c23d7 service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Refreshing network info cache for port 3fa9bd6e-092d-4e9f-a4ab-5113affb920b {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 976.561507] env[62692]: INFO nova.compute.manager [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Took 41.54 seconds to build instance. [ 976.579993] env[62692]: DEBUG nova.network.neutron [req-8080e5c8-52a8-457e-880d-09b7013d6538 req-2cc2af96-4f6f-4f10-875a-10190453eee3 service nova] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Updated VIF entry in instance network info cache for port 8842838a-68c4-4c35-b3dc-ba680c8f3732. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 976.580393] env[62692]: DEBUG nova.network.neutron [req-8080e5c8-52a8-457e-880d-09b7013d6538 req-2cc2af96-4f6f-4f10-875a-10190453eee3 service nova] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Updating instance_info_cache with network_info: [{"id": "8842838a-68c4-4c35-b3dc-ba680c8f3732", "address": "fa:16:3e:93:81:b5", "network": {"id": "1a5b25dc-fc9f-4cd6-a14c-8227afea6670", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-568391236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4d2edc0134249a590f6a25c4d372315", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8fedd232-bfc1-4e7f-bd5e-c43ef8f2f08a", "external-id": "nsx-vlan-transportzone-925", "segmentation_id": 925, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8842838a-68", "ovs_interfaceid": "8842838a-68c4-4c35-b3dc-ba680c8f3732", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.708040] env[62692]: DEBUG nova.network.neutron [-] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.708040] env[62692]: DEBUG oslo_vmware.api [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52ecf910-61a7-f561-5c8b-97b01fe875c2, 'name': SearchDatastore_Task, 'duration_secs': 0.011486} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.708743] env[62692]: DEBUG oslo_concurrency.lockutils [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.709349] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 16790505-cdf6-4937-8839-fa685a5f413e/16790505-cdf6-4937-8839-fa685a5f413e.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 976.710082] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f2d12b7e-2307-46c6-b5e2-eb9ef6943d6d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.721177] env[62692]: DEBUG oslo_vmware.api [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Waiting for the task: (returnval){ [ 976.721177] env[62692]: value = "task-1141553" [ 976.721177] env[62692]: _type = "Task" [ 976.721177] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.739870] env[62692]: DEBUG oslo_vmware.api [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Task: {'id': task-1141553, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.887469] env[62692]: DEBUG oslo_vmware.api [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141552, 'name': PowerOffVM_Task, 'duration_secs': 0.294807} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.887751] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 976.887941] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 976.888246] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-accdda8c-5c05-4544-a50c-3b2234ceda70 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.965091] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 976.965336] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 976.965546] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Deleting the datastore file [datastore2] 084ce29c-302a-45cc-b36f-0143eccee298 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 976.965923] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-585e2b56-601c-4bae-9c10-b3398ff5b04e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.975781] env[62692]: DEBUG oslo_vmware.api [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Waiting for the task: (returnval){ [ 976.975781] env[62692]: value = "task-1141555" [ 976.975781] env[62692]: _type = "Task" [ 976.975781] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.991090] env[62692]: DEBUG oslo_vmware.api [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141555, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.064287] env[62692]: DEBUG oslo_concurrency.lockutils [None req-04dd25dd-3f16-47f7-828b-8865b21e88ed tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "f7664593-1b7e-495f-8d0e-9137d7f180cc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.061s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.085689] env[62692]: DEBUG oslo_concurrency.lockutils [req-8080e5c8-52a8-457e-880d-09b7013d6538 req-2cc2af96-4f6f-4f10-875a-10190453eee3 service nova] Releasing lock "refresh_cache-16790505-cdf6-4937-8839-fa685a5f413e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.085689] env[62692]: DEBUG nova.compute.manager [req-8080e5c8-52a8-457e-880d-09b7013d6538 req-2cc2af96-4f6f-4f10-875a-10190453eee3 service nova] [instance: 46bce865-188b-487c-a73f-cf524059c797] Received event network-vif-deleted-5ff6939b-f0fd-4b3d-b0e8-f7946f80493b {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 977.086183] env[62692]: DEBUG nova.compute.manager [req-8080e5c8-52a8-457e-880d-09b7013d6538 req-2cc2af96-4f6f-4f10-875a-10190453eee3 service nova] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Received event network-vif-deleted-f1187f72-32ea-484c-9bde-f40c9a5a3a03 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 977.213067] env[62692]: INFO nova.compute.manager [-] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Took 1.03 seconds to deallocate network for instance. [ 977.238320] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Acquiring lock "f523c482-8365-47b8-9398-a56a98fcffe8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.238694] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Lock "f523c482-8365-47b8-9398-a56a98fcffe8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.240226] env[62692]: DEBUG oslo_vmware.api [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Task: {'id': task-1141553, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.367672] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e85b1903-4985-4bf4-9ed1-dd8a1eaf300b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.376190] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5be26165-27fb-45e2-892b-09e07d574291 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.416755] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3611b0e7-c9a7-4090-b8a2-3274d58cc031 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.425364] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90f04fb8-de34-44ad-b03e-de30dbba6959 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.443050] env[62692]: DEBUG nova.compute.provider_tree [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 977.488731] env[62692]: DEBUG oslo_vmware.api [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Task: {'id': task-1141555, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.385397} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.489013] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 977.489590] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 977.489681] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 977.489866] env[62692]: INFO nova.compute.manager [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Took 1.14 seconds to destroy the instance on the hypervisor. [ 977.490133] env[62692]: DEBUG oslo.service.loopingcall [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 977.490333] env[62692]: DEBUG nova.compute.manager [-] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 977.490506] env[62692]: DEBUG nova.network.neutron [-] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 977.542635] env[62692]: DEBUG nova.network.neutron [req-3bbad77a-6675-43f1-b225-47b2aa01c27b req-41a679b3-f0f6-4c05-b181-0fa2792f6618 service nova] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Updated VIF entry in instance network info cache for port 2c7422d2-dfef-4010-a99a-60c510c54a3c. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 977.543219] env[62692]: DEBUG nova.network.neutron [req-3bbad77a-6675-43f1-b225-47b2aa01c27b req-41a679b3-f0f6-4c05-b181-0fa2792f6618 service nova] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Updating instance_info_cache with network_info: [{"id": "2c7422d2-dfef-4010-a99a-60c510c54a3c", "address": "fa:16:3e:71:16:ae", "network": {"id": "17e72c3b-603f-4305-96ac-89a9b4c9c7e9", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-123501993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7db5465cb5db465b92dfa570572b6e80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c7422d2-df", "ovs_interfaceid": "2c7422d2-dfef-4010-a99a-60c510c54a3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.561036] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "f7664593-1b7e-495f-8d0e-9137d7f180cc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.561366] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "f7664593-1b7e-495f-8d0e-9137d7f180cc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.561778] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "f7664593-1b7e-495f-8d0e-9137d7f180cc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.562022] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "f7664593-1b7e-495f-8d0e-9137d7f180cc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.562209] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "f7664593-1b7e-495f-8d0e-9137d7f180cc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.564890] env[62692]: INFO nova.compute.manager [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Terminating instance [ 977.566695] env[62692]: DEBUG nova.compute.manager [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 977.567104] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 977.567949] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8721eca4-0d44-47cf-9c70-c3706f840d7b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.578151] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 977.578151] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-483e98e0-7da4-4f7d-8b0f-29da745d3aed {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.582149] env[62692]: DEBUG nova.network.neutron [req-f840c916-94c7-454c-a318-8d9ef7704ce1 req-66aca2e2-7447-48d5-af4c-5202281c23d7 service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Updated VIF entry in instance network info cache for port 3fa9bd6e-092d-4e9f-a4ab-5113affb920b. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 977.582496] env[62692]: DEBUG nova.network.neutron [req-f840c916-94c7-454c-a318-8d9ef7704ce1 req-66aca2e2-7447-48d5-af4c-5202281c23d7 service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Updating instance_info_cache with network_info: [{"id": "3fa9bd6e-092d-4e9f-a4ab-5113affb920b", "address": "fa:16:3e:a2:35:15", "network": {"id": "bbdf7bb9-6b54-4e19-91c6-cb0650e1121b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-2011930120-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "4b111e98b02a401384b23e363372ed52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5", "external-id": "nsx-vlan-transportzone-685", "segmentation_id": 685, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3fa9bd6e-09", "ovs_interfaceid": "3fa9bd6e-092d-4e9f-a4ab-5113affb920b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.586108] env[62692]: DEBUG oslo_vmware.api [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 977.586108] env[62692]: value = "task-1141556" [ 977.586108] env[62692]: _type = "Task" [ 977.586108] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.593947] env[62692]: DEBUG oslo_vmware.api [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141556, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.726974] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.734477] env[62692]: DEBUG oslo_vmware.api [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Task: {'id': task-1141553, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.572002} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.734870] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 16790505-cdf6-4937-8839-fa685a5f413e/16790505-cdf6-4937-8839-fa685a5f413e.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 977.735322] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 977.735423] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f6b64286-88fe-4eac-9420-9b6cea5b2731 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.744827] env[62692]: DEBUG nova.compute.manager [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 977.748504] env[62692]: DEBUG oslo_vmware.api [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Waiting for the task: (returnval){ [ 977.748504] env[62692]: value = "task-1141557" [ 977.748504] env[62692]: _type = "Task" [ 977.748504] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.761677] env[62692]: DEBUG oslo_vmware.api [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Task: {'id': task-1141557, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.849396] env[62692]: DEBUG nova.compute.manager [req-7c39d98d-a69a-495c-a964-84e7e1071346 req-32681832-e1c8-4635-8ef1-27928ff80385 service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Received event network-vif-deleted-3fa9bd6e-092d-4e9f-a4ab-5113affb920b {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 977.849396] env[62692]: INFO nova.compute.manager [req-7c39d98d-a69a-495c-a964-84e7e1071346 req-32681832-e1c8-4635-8ef1-27928ff80385 service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Neutron deleted interface 3fa9bd6e-092d-4e9f-a4ab-5113affb920b; detaching it from the instance and deleting it from the info cache [ 977.849396] env[62692]: DEBUG nova.network.neutron [req-7c39d98d-a69a-495c-a964-84e7e1071346 req-32681832-e1c8-4635-8ef1-27928ff80385 service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.945871] env[62692]: DEBUG nova.scheduler.client.report [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 978.048202] env[62692]: DEBUG oslo_concurrency.lockutils [req-3bbad77a-6675-43f1-b225-47b2aa01c27b req-41a679b3-f0f6-4c05-b181-0fa2792f6618 service nova] Releasing lock "refresh_cache-3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.048202] env[62692]: DEBUG nova.compute.manager [req-3bbad77a-6675-43f1-b225-47b2aa01c27b req-41a679b3-f0f6-4c05-b181-0fa2792f6618 service nova] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Received event network-changed-2c7422d2-dfef-4010-a99a-60c510c54a3c {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 978.048202] env[62692]: DEBUG nova.compute.manager [req-3bbad77a-6675-43f1-b225-47b2aa01c27b req-41a679b3-f0f6-4c05-b181-0fa2792f6618 service nova] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Refreshing instance network info cache due to event network-changed-2c7422d2-dfef-4010-a99a-60c510c54a3c. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 978.048202] env[62692]: DEBUG oslo_concurrency.lockutils [req-3bbad77a-6675-43f1-b225-47b2aa01c27b req-41a679b3-f0f6-4c05-b181-0fa2792f6618 service nova] Acquiring lock "refresh_cache-3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.048202] env[62692]: DEBUG oslo_concurrency.lockutils [req-3bbad77a-6675-43f1-b225-47b2aa01c27b req-41a679b3-f0f6-4c05-b181-0fa2792f6618 service nova] Acquired lock "refresh_cache-3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.048202] env[62692]: DEBUG nova.network.neutron [req-3bbad77a-6675-43f1-b225-47b2aa01c27b req-41a679b3-f0f6-4c05-b181-0fa2792f6618 service nova] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Refreshing network info cache for port 2c7422d2-dfef-4010-a99a-60c510c54a3c {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 978.088017] env[62692]: DEBUG oslo_concurrency.lockutils [req-f840c916-94c7-454c-a318-8d9ef7704ce1 req-66aca2e2-7447-48d5-af4c-5202281c23d7 service nova] Releasing lock "refresh_cache-084ce29c-302a-45cc-b36f-0143eccee298" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.095262] env[62692]: DEBUG oslo_vmware.api [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141556, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.262861] env[62692]: DEBUG oslo_vmware.api [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Task: {'id': task-1141557, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.321301} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.265319] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 978.266921] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5047e67-b5a4-4b5c-af3a-d1309eae8ef7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.270190] env[62692]: DEBUG nova.network.neutron [-] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.290507] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] 16790505-cdf6-4937-8839-fa685a5f413e/16790505-cdf6-4937-8839-fa685a5f413e.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 978.291782] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.292824] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-826a60e2-5778-4356-80b6-553dbba4b390 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.315903] env[62692]: DEBUG oslo_vmware.api [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Waiting for the task: (returnval){ [ 978.315903] env[62692]: value = "task-1141558" [ 978.315903] env[62692]: _type = "Task" [ 978.315903] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.326576] env[62692]: DEBUG oslo_vmware.api [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Task: {'id': task-1141558, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.352323] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7f0c57d7-1b0d-4ad2-9e5b-ea5aae335716 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.362490] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01931b71-29f3-4122-b0c5-3bdb847c4122 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.397854] env[62692]: DEBUG nova.compute.manager [req-7c39d98d-a69a-495c-a964-84e7e1071346 req-32681832-e1c8-4635-8ef1-27928ff80385 service nova] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Detach interface failed, port_id=3fa9bd6e-092d-4e9f-a4ab-5113affb920b, reason: Instance 084ce29c-302a-45cc-b36f-0143eccee298 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 978.454703] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.106s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.460135] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 25.708s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.478020] env[62692]: INFO nova.scheduler.client.report [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Deleted allocations for instance e2b3b062-10e2-4a9c-b6b3-f0786de07b79 [ 978.596568] env[62692]: DEBUG oslo_vmware.api [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141556, 'name': PowerOffVM_Task, 'duration_secs': 0.598635} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.596869] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 978.602020] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 978.602020] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-17302a89-9bc8-4725-8707-8739675c5ee0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.793166] env[62692]: INFO nova.compute.manager [-] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Took 1.30 seconds to deallocate network for instance. [ 978.825883] env[62692]: DEBUG oslo_vmware.api [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Task: {'id': task-1141558, 'name': ReconfigVM_Task, 'duration_secs': 0.337354} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.826375] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Reconfigured VM instance instance-00000052 to attach disk [datastore1] 16790505-cdf6-4937-8839-fa685a5f413e/16790505-cdf6-4937-8839-fa685a5f413e.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 978.827190] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3b95a3d8-b0e0-418e-8d25-6114b03c94d8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.840023] env[62692]: DEBUG oslo_vmware.api [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Waiting for the task: (returnval){ [ 978.840023] env[62692]: value = "task-1141560" [ 978.840023] env[62692]: _type = "Task" [ 978.840023] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.853894] env[62692]: DEBUG oslo_vmware.api [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Task: {'id': task-1141560, 'name': Rename_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.964482] env[62692]: INFO nova.compute.claims [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 978.990168] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e5172114-08da-47fb-856d-899ecf0d9331 tempest-ServerTagsTestJSON-1910346210 tempest-ServerTagsTestJSON-1910346210-project-member] Lock "e2b3b062-10e2-4a9c-b6b3-f0786de07b79" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.411s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.086326] env[62692]: DEBUG nova.network.neutron [req-3bbad77a-6675-43f1-b225-47b2aa01c27b req-41a679b3-f0f6-4c05-b181-0fa2792f6618 service nova] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Updated VIF entry in instance network info cache for port 2c7422d2-dfef-4010-a99a-60c510c54a3c. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 979.086734] env[62692]: DEBUG nova.network.neutron [req-3bbad77a-6675-43f1-b225-47b2aa01c27b req-41a679b3-f0f6-4c05-b181-0fa2792f6618 service nova] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Updating instance_info_cache with network_info: [{"id": "2c7422d2-dfef-4010-a99a-60c510c54a3c", "address": "fa:16:3e:71:16:ae", "network": {"id": "17e72c3b-603f-4305-96ac-89a9b4c9c7e9", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-123501993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7db5465cb5db465b92dfa570572b6e80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c7422d2-df", "ovs_interfaceid": "2c7422d2-dfef-4010-a99a-60c510c54a3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.300166] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.347782] env[62692]: DEBUG oslo_vmware.api [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Task: {'id': task-1141560, 'name': Rename_Task, 'duration_secs': 0.166995} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.348112] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 979.348368] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9d565721-6b72-4b08-b998-1014c5a849bf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.355995] env[62692]: DEBUG oslo_vmware.api [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Waiting for the task: (returnval){ [ 979.355995] env[62692]: value = "task-1141561" [ 979.355995] env[62692]: _type = "Task" [ 979.355995] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.365182] env[62692]: DEBUG oslo_vmware.api [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Task: {'id': task-1141561, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.471686] env[62692]: INFO nova.compute.resource_tracker [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Updating resource usage from migration 4c7b8a13-dd0a-48d9-b1fe-523c53945e0c [ 979.592916] env[62692]: DEBUG oslo_concurrency.lockutils [req-3bbad77a-6675-43f1-b225-47b2aa01c27b req-41a679b3-f0f6-4c05-b181-0fa2792f6618 service nova] Releasing lock "refresh_cache-3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.862838] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a10114c-443c-41d1-bcc9-e2a50b18a484 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.870402] env[62692]: DEBUG oslo_vmware.api [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Task: {'id': task-1141561, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.874462] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c543ee84-4e73-4bb6-9687-864c394b7a37 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.909697] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f21914ec-0b79-48a9-b0c2-473bb6ced32b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.916520] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe936485-d3e5-4838-a8dc-ef9b974b8159 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.934519] env[62692]: DEBUG nova.compute.provider_tree [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.367498] env[62692]: DEBUG oslo_vmware.api [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Task: {'id': task-1141561, 'name': PowerOnVM_Task, 'duration_secs': 0.909576} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.367757] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 980.367959] env[62692]: INFO nova.compute.manager [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Took 9.35 seconds to spawn the instance on the hypervisor. [ 980.368247] env[62692]: DEBUG nova.compute.manager [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 980.369026] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc065468-6697-4e2d-888f-e66f3f1fa43b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.437278] env[62692]: DEBUG nova.scheduler.client.report [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 980.891244] env[62692]: INFO nova.compute.manager [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Took 43.94 seconds to build instance. [ 980.943030] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.486s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.943258] env[62692]: INFO nova.compute.manager [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Migrating [ 980.953694] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 23.406s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.953921] env[62692]: DEBUG nova.objects.instance [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62692) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 981.369525] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 981.369525] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 981.369525] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Deleting the datastore file [datastore1] f7664593-1b7e-495f-8d0e-9137d7f180cc {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 981.369525] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d826cd8e-58b3-4c8c-a46d-c9ad1778c9dd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.377193] env[62692]: DEBUG oslo_vmware.api [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 981.377193] env[62692]: value = "task-1141562" [ 981.377193] env[62692]: _type = "Task" [ 981.377193] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.388369] env[62692]: DEBUG oslo_vmware.api [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141562, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.395415] env[62692]: DEBUG oslo_concurrency.lockutils [None req-069e9fe6-73f7-44c7-a321-623b74464c57 tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Lock "16790505-cdf6-4937-8839-fa685a5f413e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.449s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.468756] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "refresh_cache-3f6c9744-a6e3-43f3-8b6b-624ba178747d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.470635] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired lock "refresh_cache-3f6c9744-a6e3-43f3-8b6b-624ba178747d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.470635] env[62692]: DEBUG nova.network.neutron [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 981.757876] env[62692]: DEBUG nova.compute.manager [req-5e0637ff-6991-4bba-88d3-5bbbebd3d3b7 req-a30e3d29-d9c2-4c72-9fdb-873cc8df36b7 service nova] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Received event network-changed-8842838a-68c4-4c35-b3dc-ba680c8f3732 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 981.758121] env[62692]: DEBUG nova.compute.manager [req-5e0637ff-6991-4bba-88d3-5bbbebd3d3b7 req-a30e3d29-d9c2-4c72-9fdb-873cc8df36b7 service nova] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Refreshing instance network info cache due to event network-changed-8842838a-68c4-4c35-b3dc-ba680c8f3732. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 981.758323] env[62692]: DEBUG oslo_concurrency.lockutils [req-5e0637ff-6991-4bba-88d3-5bbbebd3d3b7 req-a30e3d29-d9c2-4c72-9fdb-873cc8df36b7 service nova] Acquiring lock "refresh_cache-16790505-cdf6-4937-8839-fa685a5f413e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.758472] env[62692]: DEBUG oslo_concurrency.lockutils [req-5e0637ff-6991-4bba-88d3-5bbbebd3d3b7 req-a30e3d29-d9c2-4c72-9fdb-873cc8df36b7 service nova] Acquired lock "refresh_cache-16790505-cdf6-4937-8839-fa685a5f413e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.758637] env[62692]: DEBUG nova.network.neutron [req-5e0637ff-6991-4bba-88d3-5bbbebd3d3b7 req-a30e3d29-d9c2-4c72-9fdb-873cc8df36b7 service nova] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Refreshing network info cache for port 8842838a-68c4-4c35-b3dc-ba680c8f3732 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 981.887562] env[62692]: DEBUG oslo_vmware.api [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141562, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165169} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.887869] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 981.888086] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 981.888303] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 981.888564] env[62692]: INFO nova.compute.manager [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Took 4.32 seconds to destroy the instance on the hypervisor. [ 981.888702] env[62692]: DEBUG oslo.service.loopingcall [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 981.888964] env[62692]: DEBUG nova.compute.manager [-] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 981.889017] env[62692]: DEBUG nova.network.neutron [-] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 981.965729] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ef8c7f77-6cdf-4eee-9872-d3a498b34a40 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.966925] env[62692]: DEBUG oslo_concurrency.lockutils [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.406s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.967255] env[62692]: DEBUG nova.objects.instance [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lazy-loading 'resources' on Instance uuid 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 982.236951] env[62692]: DEBUG nova.network.neutron [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Updating instance_info_cache with network_info: [{"id": "2cb4b52c-0d89-4402-ae48-dc6f226c557c", "address": "fa:16:3e:c8:b1:41", "network": {"id": "f6352999-f943-431d-b54a-c9dc1d810a85", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-121037111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9cb8ae08824245a7911de7455e7fcb68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c842425c-544e-4ce2-9657-512723bd318e", "external-id": "nsx-vlan-transportzone-80", "segmentation_id": 80, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cb4b52c-0d", "ovs_interfaceid": "2cb4b52c-0d89-4402-ae48-dc6f226c557c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.692793] env[62692]: DEBUG nova.network.neutron [req-5e0637ff-6991-4bba-88d3-5bbbebd3d3b7 req-a30e3d29-d9c2-4c72-9fdb-873cc8df36b7 service nova] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Updated VIF entry in instance network info cache for port 8842838a-68c4-4c35-b3dc-ba680c8f3732. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 982.692793] env[62692]: DEBUG nova.network.neutron [req-5e0637ff-6991-4bba-88d3-5bbbebd3d3b7 req-a30e3d29-d9c2-4c72-9fdb-873cc8df36b7 service nova] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Updating instance_info_cache with network_info: [{"id": "8842838a-68c4-4c35-b3dc-ba680c8f3732", "address": "fa:16:3e:93:81:b5", "network": {"id": "1a5b25dc-fc9f-4cd6-a14c-8227afea6670", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-568391236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.153", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4d2edc0134249a590f6a25c4d372315", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8fedd232-bfc1-4e7f-bd5e-c43ef8f2f08a", "external-id": "nsx-vlan-transportzone-925", "segmentation_id": 925, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8842838a-68", "ovs_interfaceid": "8842838a-68c4-4c35-b3dc-ba680c8f3732", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.740349] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Releasing lock "refresh_cache-3f6c9744-a6e3-43f3-8b6b-624ba178747d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.776473] env[62692]: DEBUG nova.network.neutron [-] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.888301] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2a159dd-91d7-43d2-87cf-a16d96e7ab91 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.898630] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de4b7aa6-fc6f-47a1-aa8c-7d9c7ca4fa37 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.935954] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47245e50-5bd9-4908-baae-743024482f32 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.945487] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49e15463-7320-42ef-9145-9fceb02b5543 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.962697] env[62692]: DEBUG nova.compute.provider_tree [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 983.196147] env[62692]: DEBUG oslo_concurrency.lockutils [req-5e0637ff-6991-4bba-88d3-5bbbebd3d3b7 req-a30e3d29-d9c2-4c72-9fdb-873cc8df36b7 service nova] Releasing lock "refresh_cache-16790505-cdf6-4937-8839-fa685a5f413e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.282893] env[62692]: INFO nova.compute.manager [-] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Took 1.39 seconds to deallocate network for instance. [ 983.465792] env[62692]: DEBUG nova.scheduler.client.report [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 983.790202] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.800223] env[62692]: DEBUG nova.compute.manager [req-fa215072-2321-4c93-a375-7ca005f87cb9 req-a50c0d67-67b2-4080-ae15-28229ef79394 service nova] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Received event network-vif-deleted-daea3ee8-3a77-43be-b648-99133117b287 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 983.973770] env[62692]: DEBUG oslo_concurrency.lockutils [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.007s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.976985] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 17.807s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.976985] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.976985] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62692) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 983.977418] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.960s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.978609] env[62692]: INFO nova.compute.claims [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 983.982334] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-742ad7cb-1010-4b07-8658-13fe0797d5b0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.993238] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ba40843-c3d2-45f7-a91b-c16cc6ec3432 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.998373] env[62692]: INFO nova.scheduler.client.report [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Deleted allocations for instance 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d [ 984.010604] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68591448-7206-4ed9-857f-269b3fbf5f58 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.018628] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-042afa18-d667-45ec-af01-75b15f89248f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.052814] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=178741MB free_disk=165GB free_vcpus=48 pci_devices=None {{(pid=62692) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 984.053040] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.260895] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acfbec71-ee6e-4c8f-91f5-2d0fb1d8a4a4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.282837] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Updating instance '3f6c9744-a6e3-43f3-8b6b-624ba178747d' progress to 0 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 984.513323] env[62692]: DEBUG oslo_concurrency.lockutils [None req-493ede69-e8ff-4342-a11b-7db0d1bcb95a tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.962s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.789846] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 984.790331] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c05826ec-0871-4605-9e5d-bf5e53f276b7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.799780] env[62692]: DEBUG oslo_vmware.api [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 984.799780] env[62692]: value = "task-1141563" [ 984.799780] env[62692]: _type = "Task" [ 984.799780] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.809873] env[62692]: DEBUG oslo_vmware.api [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141563, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.905134] env[62692]: DEBUG oslo_concurrency.lockutils [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "26c1ff5b-9eda-4131-a4ea-d4511a2364b7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.905421] env[62692]: DEBUG oslo_concurrency.lockutils [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "26c1ff5b-9eda-4131-a4ea-d4511a2364b7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.905994] env[62692]: DEBUG oslo_concurrency.lockutils [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "26c1ff5b-9eda-4131-a4ea-d4511a2364b7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.906308] env[62692]: DEBUG oslo_concurrency.lockutils [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "26c1ff5b-9eda-4131-a4ea-d4511a2364b7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.906587] env[62692]: DEBUG oslo_concurrency.lockutils [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "26c1ff5b-9eda-4131-a4ea-d4511a2364b7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.909601] env[62692]: INFO nova.compute.manager [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Terminating instance [ 984.913325] env[62692]: DEBUG nova.compute.manager [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 984.913325] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 984.913325] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf6eb7db-dc72-4ce0-87ec-9899a645d1b4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.924858] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 984.924858] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b5bfb52c-9af6-431e-9c7f-4cba48f0cf66 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.932451] env[62692]: DEBUG oslo_vmware.api [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 984.932451] env[62692]: value = "task-1141564" [ 984.932451] env[62692]: _type = "Task" [ 984.932451] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.943522] env[62692]: DEBUG oslo_vmware.api [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141564, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.309823] env[62692]: DEBUG oslo_vmware.api [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141563, 'name': PowerOffVM_Task, 'duration_secs': 0.209079} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.310340] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 985.310384] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Updating instance '3f6c9744-a6e3-43f3-8b6b-624ba178747d' progress to 17 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 985.433298] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0577d561-a54e-4084-907e-ff2df290a8ae {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.440444] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquiring lock "1b753882-9118-479c-aa3e-cbb8bdf2e086" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.440689] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Lock "1b753882-9118-479c-aa3e-cbb8bdf2e086" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.453124] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-378c91c6-8fa8-4b48-8902-0c76e9dcccc7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.456415] env[62692]: DEBUG oslo_vmware.api [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141564, 'name': PowerOffVM_Task, 'duration_secs': 0.274297} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.457173] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 985.457366] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 985.458157] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-240cd7fe-0bb3-4bc2-ab3a-652b263627ad {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.488534] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5460b15-bfff-4312-a238-136af4139456 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.498849] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26d924be-d1de-46b7-98b2-a41a6e924c36 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.514133] env[62692]: DEBUG nova.compute.provider_tree [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 985.544164] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 985.544420] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 985.545143] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Deleting the datastore file [datastore1] 26c1ff5b-9eda-4131-a4ea-d4511a2364b7 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 985.545143] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-83f8510e-f604-4a64-bd8a-6be8bce285d7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.552113] env[62692]: DEBUG oslo_vmware.api [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 985.552113] env[62692]: value = "task-1141566" [ 985.552113] env[62692]: _type = "Task" [ 985.552113] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.817546] env[62692]: DEBUG nova.virt.hardware [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 985.817762] env[62692]: DEBUG nova.virt.hardware [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 985.817963] env[62692]: DEBUG nova.virt.hardware [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 985.818197] env[62692]: DEBUG nova.virt.hardware [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 985.818355] env[62692]: DEBUG nova.virt.hardware [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 985.818513] env[62692]: DEBUG nova.virt.hardware [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 985.818717] env[62692]: DEBUG nova.virt.hardware [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 985.818880] env[62692]: DEBUG nova.virt.hardware [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 985.819081] env[62692]: DEBUG nova.virt.hardware [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 985.819352] env[62692]: DEBUG nova.virt.hardware [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 985.819573] env[62692]: DEBUG nova.virt.hardware [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 985.824706] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ddb2615-885c-487d-aa2c-1183c035c800 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.842886] env[62692]: DEBUG oslo_vmware.api [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 985.842886] env[62692]: value = "task-1141567" [ 985.842886] env[62692]: _type = "Task" [ 985.842886] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.851783] env[62692]: DEBUG oslo_vmware.api [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141567, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.946441] env[62692]: DEBUG nova.compute.manager [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 986.017815] env[62692]: DEBUG nova.scheduler.client.report [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 986.068747] env[62692]: DEBUG oslo_vmware.api [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141566, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.181792} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.069944] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 986.070263] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 986.071338] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 986.071658] env[62692]: INFO nova.compute.manager [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Took 1.16 seconds to destroy the instance on the hypervisor. [ 986.072020] env[62692]: DEBUG oslo.service.loopingcall [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 986.072347] env[62692]: DEBUG nova.compute.manager [-] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 986.072533] env[62692]: DEBUG nova.network.neutron [-] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 986.354466] env[62692]: DEBUG oslo_vmware.api [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141567, 'name': ReconfigVM_Task, 'duration_secs': 0.222719} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.354889] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Updating instance '3f6c9744-a6e3-43f3-8b6b-624ba178747d' progress to 33 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 986.362184] env[62692]: DEBUG nova.compute.manager [req-e6dc1251-a3e4-4c03-a3b5-fec186703e8b req-57b40c4f-6085-4f3e-a984-00ac207884f0 service nova] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Received event network-vif-deleted-513b63b4-6498-4af7-b62f-9bdbc4f1e9ed {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 986.362184] env[62692]: INFO nova.compute.manager [req-e6dc1251-a3e4-4c03-a3b5-fec186703e8b req-57b40c4f-6085-4f3e-a984-00ac207884f0 service nova] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Neutron deleted interface 513b63b4-6498-4af7-b62f-9bdbc4f1e9ed; detaching it from the instance and deleting it from the info cache [ 986.362184] env[62692]: DEBUG nova.network.neutron [req-e6dc1251-a3e4-4c03-a3b5-fec186703e8b req-57b40c4f-6085-4f3e-a984-00ac207884f0 service nova] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.470355] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.522207] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.545s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.522789] env[62692]: DEBUG nova.compute.manager [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 986.525381] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.366s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.526839] env[62692]: INFO nova.compute.claims [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 986.836127] env[62692]: DEBUG nova.network.neutron [-] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.869239] env[62692]: DEBUG nova.virt.hardware [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 986.869483] env[62692]: DEBUG nova.virt.hardware [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 986.869659] env[62692]: DEBUG nova.virt.hardware [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 986.869846] env[62692]: DEBUG nova.virt.hardware [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 986.869997] env[62692]: DEBUG nova.virt.hardware [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 986.870166] env[62692]: DEBUG nova.virt.hardware [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 986.870422] env[62692]: DEBUG nova.virt.hardware [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 986.870614] env[62692]: DEBUG nova.virt.hardware [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 986.870793] env[62692]: DEBUG nova.virt.hardware [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 986.870962] env[62692]: DEBUG nova.virt.hardware [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 986.871161] env[62692]: DEBUG nova.virt.hardware [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 986.876357] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Reconfiguring VM instance instance-0000004b to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 986.876857] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6d734bf5-5b51-4149-84e3-725f27e35f5e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.890064] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3dc5f19c-4f18-49d9-a373-b74dbc708ae1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.901752] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95dfd616-7c74-4225-be92-392a09923306 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.912136] env[62692]: DEBUG oslo_vmware.api [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 986.912136] env[62692]: value = "task-1141568" [ 986.912136] env[62692]: _type = "Task" [ 986.912136] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.920120] env[62692]: DEBUG oslo_vmware.api [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141568, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.935356] env[62692]: DEBUG nova.compute.manager [req-e6dc1251-a3e4-4c03-a3b5-fec186703e8b req-57b40c4f-6085-4f3e-a984-00ac207884f0 service nova] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Detach interface failed, port_id=513b63b4-6498-4af7-b62f-9bdbc4f1e9ed, reason: Instance 26c1ff5b-9eda-4131-a4ea-d4511a2364b7 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 987.031706] env[62692]: DEBUG nova.compute.utils [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 987.035720] env[62692]: DEBUG nova.compute.manager [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 987.035720] env[62692]: DEBUG nova.network.neutron [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 987.095032] env[62692]: DEBUG nova.policy [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a28f5e0e0af443c5be82214d446c221d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '589895ae20394e2fa525707bf7d09f2a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 987.340704] env[62692]: INFO nova.compute.manager [-] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Took 1.27 seconds to deallocate network for instance. [ 987.370834] env[62692]: DEBUG nova.network.neutron [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Successfully created port: 30d6694e-3be7-4c78-96c8-c9e2005856a5 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 987.423220] env[62692]: DEBUG oslo_vmware.api [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141568, 'name': ReconfigVM_Task, 'duration_secs': 0.16769} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.424061] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Reconfigured VM instance instance-0000004b to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 987.425399] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cab85052-f4df-48e0-8009-ff2f5c144771 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.457240] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 3f6c9744-a6e3-43f3-8b6b-624ba178747d/3f6c9744-a6e3-43f3-8b6b-624ba178747d.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 987.458149] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ba75610-6ecd-48c7-b5c1-5644e571f099 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.481955] env[62692]: DEBUG oslo_vmware.api [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 987.481955] env[62692]: value = "task-1141569" [ 987.481955] env[62692]: _type = "Task" [ 987.481955] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.487528] env[62692]: DEBUG oslo_vmware.api [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141569, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.539019] env[62692]: DEBUG nova.compute.manager [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 987.849062] env[62692]: DEBUG oslo_concurrency.lockutils [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.879192] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a54b9eb-5ddf-4fa1-b43e-850caaa9b52f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.887036] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1baf57b3-36f8-41d5-a9bd-3ef36af45c7b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.918810] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d8372db-96a7-47ea-9026-74233aebad30 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.927018] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbd0792b-a271-4b2e-babd-384f09b627bc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.944082] env[62692]: DEBUG nova.compute.provider_tree [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 987.988456] env[62692]: DEBUG oslo_vmware.api [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141569, 'name': ReconfigVM_Task, 'duration_secs': 0.261215} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.988799] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 3f6c9744-a6e3-43f3-8b6b-624ba178747d/3f6c9744-a6e3-43f3-8b6b-624ba178747d.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 987.989132] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Updating instance '3f6c9744-a6e3-43f3-8b6b-624ba178747d' progress to 50 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 988.446921] env[62692]: DEBUG nova.scheduler.client.report [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 988.495354] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b35b7e7-8fce-48a4-811a-92ff7b2daa12 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.515283] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db1f3e10-f89d-42e2-81c1-36d8da2b1459 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.534309] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Updating instance '3f6c9744-a6e3-43f3-8b6b-624ba178747d' progress to 67 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 988.548683] env[62692]: DEBUG nova.compute.manager [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 988.573768] env[62692]: DEBUG nova.virt.hardware [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 988.574110] env[62692]: DEBUG nova.virt.hardware [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 988.574321] env[62692]: DEBUG nova.virt.hardware [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 988.574555] env[62692]: DEBUG nova.virt.hardware [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 988.574767] env[62692]: DEBUG nova.virt.hardware [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 988.574970] env[62692]: DEBUG nova.virt.hardware [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 988.575212] env[62692]: DEBUG nova.virt.hardware [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 988.575381] env[62692]: DEBUG nova.virt.hardware [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 988.575555] env[62692]: DEBUG nova.virt.hardware [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 988.575720] env[62692]: DEBUG nova.virt.hardware [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 988.575899] env[62692]: DEBUG nova.virt.hardware [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 988.576761] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8abc881f-ee00-4277-9bf9-133752a0db74 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.585589] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a45e9d0-ff56-4f45-85b1-03ba3aef9f17 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.788752] env[62692]: DEBUG nova.compute.manager [req-99d1d087-596f-4def-b75b-8398bb451572 req-36cd52e9-a103-4dcc-b2b0-642c2c38207b service nova] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Received event network-vif-plugged-30d6694e-3be7-4c78-96c8-c9e2005856a5 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 988.789053] env[62692]: DEBUG oslo_concurrency.lockutils [req-99d1d087-596f-4def-b75b-8398bb451572 req-36cd52e9-a103-4dcc-b2b0-642c2c38207b service nova] Acquiring lock "250b9cc2-706e-4fbf-887c-1c80fc27a0ed-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.789367] env[62692]: DEBUG oslo_concurrency.lockutils [req-99d1d087-596f-4def-b75b-8398bb451572 req-36cd52e9-a103-4dcc-b2b0-642c2c38207b service nova] Lock "250b9cc2-706e-4fbf-887c-1c80fc27a0ed-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.789550] env[62692]: DEBUG oslo_concurrency.lockutils [req-99d1d087-596f-4def-b75b-8398bb451572 req-36cd52e9-a103-4dcc-b2b0-642c2c38207b service nova] Lock "250b9cc2-706e-4fbf-887c-1c80fc27a0ed-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.789770] env[62692]: DEBUG nova.compute.manager [req-99d1d087-596f-4def-b75b-8398bb451572 req-36cd52e9-a103-4dcc-b2b0-642c2c38207b service nova] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] No waiting events found dispatching network-vif-plugged-30d6694e-3be7-4c78-96c8-c9e2005856a5 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 988.789967] env[62692]: WARNING nova.compute.manager [req-99d1d087-596f-4def-b75b-8398bb451572 req-36cd52e9-a103-4dcc-b2b0-642c2c38207b service nova] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Received unexpected event network-vif-plugged-30d6694e-3be7-4c78-96c8-c9e2005856a5 for instance with vm_state building and task_state spawning. [ 988.884126] env[62692]: DEBUG nova.network.neutron [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Successfully updated port: 30d6694e-3be7-4c78-96c8-c9e2005856a5 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 988.952577] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.427s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.953144] env[62692]: DEBUG nova.compute.manager [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 988.956305] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 19.819s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.956503] env[62692]: DEBUG nova.objects.instance [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 46bce865-188b-487c-a73f-cf524059c797] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62692) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 989.280093] env[62692]: DEBUG nova.network.neutron [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Port 2cb4b52c-0d89-4402-ae48-dc6f226c557c binding to destination host cpu-1 is already ACTIVE {{(pid=62692) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 989.387792] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "refresh_cache-250b9cc2-706e-4fbf-887c-1c80fc27a0ed" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.387944] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired lock "refresh_cache-250b9cc2-706e-4fbf-887c-1c80fc27a0ed" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.388117] env[62692]: DEBUG nova.network.neutron [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 989.464746] env[62692]: DEBUG nova.compute.utils [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 989.465536] env[62692]: DEBUG nova.compute.manager [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 989.465710] env[62692]: DEBUG nova.network.neutron [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 989.505778] env[62692]: DEBUG nova.policy [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '47c402e4831b4ca6b7db735d109523de', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eb2ee16ec90c48e6babfb1d31ea3de94', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 989.822794] env[62692]: DEBUG nova.network.neutron [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Successfully created port: ab9d5209-bc8f-437d-a219-7f521cae10e6 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 989.928580] env[62692]: DEBUG nova.network.neutron [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 989.968869] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cede118a-7e8f-41b2-9698-ef66e3b49695 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.971055] env[62692]: DEBUG nova.compute.manager [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 989.974051] env[62692]: DEBUG oslo_concurrency.lockutils [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 19.838s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.974320] env[62692]: DEBUG nova.objects.instance [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62692) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 990.089047] env[62692]: DEBUG nova.network.neutron [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Updating instance_info_cache with network_info: [{"id": "30d6694e-3be7-4c78-96c8-c9e2005856a5", "address": "fa:16:3e:b5:ab:9f", "network": {"id": "dd260020-2d74-46ba-aa78-af29bdefa176", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-617179655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "589895ae20394e2fa525707bf7d09f2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30d6694e-3b", "ovs_interfaceid": "30d6694e-3be7-4c78-96c8-c9e2005856a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.300601] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "3f6c9744-a6e3-43f3-8b6b-624ba178747d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.300844] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "3f6c9744-a6e3-43f3-8b6b-624ba178747d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.301036] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "3f6c9744-a6e3-43f3-8b6b-624ba178747d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.591335] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Releasing lock "refresh_cache-250b9cc2-706e-4fbf-887c-1c80fc27a0ed" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.591697] env[62692]: DEBUG nova.compute.manager [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Instance network_info: |[{"id": "30d6694e-3be7-4c78-96c8-c9e2005856a5", "address": "fa:16:3e:b5:ab:9f", "network": {"id": "dd260020-2d74-46ba-aa78-af29bdefa176", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-617179655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "589895ae20394e2fa525707bf7d09f2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30d6694e-3b", "ovs_interfaceid": "30d6694e-3be7-4c78-96c8-c9e2005856a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 990.592145] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b5:ab:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55764410-260e-4339-a020-6b30995584bf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '30d6694e-3be7-4c78-96c8-c9e2005856a5', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 990.599815] env[62692]: DEBUG oslo.service.loopingcall [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 990.600031] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 990.600303] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d324e3e1-fcf5-462b-afff-dc2fa2f664c7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.622802] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 990.622802] env[62692]: value = "task-1141570" [ 990.622802] env[62692]: _type = "Task" [ 990.622802] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.630615] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141570, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.817057] env[62692]: DEBUG nova.compute.manager [req-e7894d5a-b5b3-47ac-8b1e-ff9db9df32df req-f24c1bf4-7fcf-452f-bf28-2f74a5243a36 service nova] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Received event network-changed-30d6694e-3be7-4c78-96c8-c9e2005856a5 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 990.817343] env[62692]: DEBUG nova.compute.manager [req-e7894d5a-b5b3-47ac-8b1e-ff9db9df32df req-f24c1bf4-7fcf-452f-bf28-2f74a5243a36 service nova] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Refreshing instance network info cache due to event network-changed-30d6694e-3be7-4c78-96c8-c9e2005856a5. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 990.817584] env[62692]: DEBUG oslo_concurrency.lockutils [req-e7894d5a-b5b3-47ac-8b1e-ff9db9df32df req-f24c1bf4-7fcf-452f-bf28-2f74a5243a36 service nova] Acquiring lock "refresh_cache-250b9cc2-706e-4fbf-887c-1c80fc27a0ed" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.817748] env[62692]: DEBUG oslo_concurrency.lockutils [req-e7894d5a-b5b3-47ac-8b1e-ff9db9df32df req-f24c1bf4-7fcf-452f-bf28-2f74a5243a36 service nova] Acquired lock "refresh_cache-250b9cc2-706e-4fbf-887c-1c80fc27a0ed" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.817921] env[62692]: DEBUG nova.network.neutron [req-e7894d5a-b5b3-47ac-8b1e-ff9db9df32df req-f24c1bf4-7fcf-452f-bf28-2f74a5243a36 service nova] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Refreshing network info cache for port 30d6694e-3be7-4c78-96c8-c9e2005856a5 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 990.986747] env[62692]: DEBUG nova.compute.manager [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 990.989594] env[62692]: DEBUG oslo_concurrency.lockutils [None req-53ca5fb3-7b76-4326-a0a9-c70b6daf1be8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.991099] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.654s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.992309] env[62692]: INFO nova.compute.claims [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 991.015238] env[62692]: DEBUG nova.virt.hardware [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 991.015504] env[62692]: DEBUG nova.virt.hardware [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 991.015670] env[62692]: DEBUG nova.virt.hardware [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 991.015874] env[62692]: DEBUG nova.virt.hardware [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 991.016055] env[62692]: DEBUG nova.virt.hardware [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 991.016245] env[62692]: DEBUG nova.virt.hardware [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 991.016731] env[62692]: DEBUG nova.virt.hardware [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 991.016731] env[62692]: DEBUG nova.virt.hardware [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 991.016851] env[62692]: DEBUG nova.virt.hardware [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 991.017038] env[62692]: DEBUG nova.virt.hardware [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 991.017203] env[62692]: DEBUG nova.virt.hardware [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 991.018114] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68af26a9-f5c3-451a-a1ea-cd77026b7eb5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.026904] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e968135-d6ea-40c7-b9ea-180aa72e30f0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.133427] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141570, 'name': CreateVM_Task, 'duration_secs': 0.346899} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.133427] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 991.134116] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.134294] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.134687] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 991.134968] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95faef5f-0619-4c4d-b27f-d7c517d481df {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.140244] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 991.140244] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]528c1a1e-f4f4-deef-a66f-1a55f7f64b60" [ 991.140244] env[62692]: _type = "Task" [ 991.140244] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.148058] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]528c1a1e-f4f4-deef-a66f-1a55f7f64b60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.309810] env[62692]: DEBUG nova.network.neutron [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Successfully updated port: ab9d5209-bc8f-437d-a219-7f521cae10e6 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 991.360575] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "refresh_cache-3f6c9744-a6e3-43f3-8b6b-624ba178747d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.360575] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired lock "refresh_cache-3f6c9744-a6e3-43f3-8b6b-624ba178747d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.360575] env[62692]: DEBUG nova.network.neutron [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 991.548992] env[62692]: DEBUG nova.network.neutron [req-e7894d5a-b5b3-47ac-8b1e-ff9db9df32df req-f24c1bf4-7fcf-452f-bf28-2f74a5243a36 service nova] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Updated VIF entry in instance network info cache for port 30d6694e-3be7-4c78-96c8-c9e2005856a5. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 991.549443] env[62692]: DEBUG nova.network.neutron [req-e7894d5a-b5b3-47ac-8b1e-ff9db9df32df req-f24c1bf4-7fcf-452f-bf28-2f74a5243a36 service nova] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Updating instance_info_cache with network_info: [{"id": "30d6694e-3be7-4c78-96c8-c9e2005856a5", "address": "fa:16:3e:b5:ab:9f", "network": {"id": "dd260020-2d74-46ba-aa78-af29bdefa176", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-617179655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "589895ae20394e2fa525707bf7d09f2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30d6694e-3b", "ovs_interfaceid": "30d6694e-3be7-4c78-96c8-c9e2005856a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.651121] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]528c1a1e-f4f4-deef-a66f-1a55f7f64b60, 'name': SearchDatastore_Task, 'duration_secs': 0.013367} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.651457] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.651698] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 991.651975] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.652088] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.652271] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 991.652547] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-26ac4e31-31b3-452e-b0be-f0651a43abcf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.661782] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 991.661892] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 991.662592] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03ab8f0a-8f15-4d02-a967-1e46de819d27 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.667836] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 991.667836] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]520b3d08-0ac9-2dfe-43fc-e4bb9caba2b6" [ 991.667836] env[62692]: _type = "Task" [ 991.667836] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.674778] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]520b3d08-0ac9-2dfe-43fc-e4bb9caba2b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.812159] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "refresh_cache-7183ec4a-992d-48f1-8dda-7f499c2f4e1d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.812323] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquired lock "refresh_cache-7183ec4a-992d-48f1-8dda-7f499c2f4e1d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.812501] env[62692]: DEBUG nova.network.neutron [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 992.052527] env[62692]: DEBUG oslo_concurrency.lockutils [req-e7894d5a-b5b3-47ac-8b1e-ff9db9df32df req-f24c1bf4-7fcf-452f-bf28-2f74a5243a36 service nova] Releasing lock "refresh_cache-250b9cc2-706e-4fbf-887c-1c80fc27a0ed" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.097183] env[62692]: DEBUG nova.network.neutron [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Updating instance_info_cache with network_info: [{"id": "2cb4b52c-0d89-4402-ae48-dc6f226c557c", "address": "fa:16:3e:c8:b1:41", "network": {"id": "f6352999-f943-431d-b54a-c9dc1d810a85", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-121037111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9cb8ae08824245a7911de7455e7fcb68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c842425c-544e-4ce2-9657-512723bd318e", "external-id": "nsx-vlan-transportzone-80", "segmentation_id": 80, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cb4b52c-0d", "ovs_interfaceid": "2cb4b52c-0d89-4402-ae48-dc6f226c557c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.178259] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]520b3d08-0ac9-2dfe-43fc-e4bb9caba2b6, 'name': SearchDatastore_Task, 'duration_secs': 0.008646} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.181162] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66c7970d-26cd-48ab-b8d5-13ec810c0855 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.186006] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 992.186006] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]522eaf2c-ff18-a3fa-476f-9adedbb2b66c" [ 992.186006] env[62692]: _type = "Task" [ 992.186006] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.195626] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]522eaf2c-ff18-a3fa-476f-9adedbb2b66c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.296460] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90f300ba-c7c6-46a9-9cef-e3e23953e4ec {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.303666] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d56f0fde-8e90-4b8a-8ea0-6428165cecf8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.334036] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54359b9a-0f59-42af-9e3f-45016b6b43f3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.341013] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d36dbed5-ee1b-48dd-90ef-3118fa34dcfb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.355559] env[62692]: DEBUG nova.compute.provider_tree [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.363699] env[62692]: DEBUG nova.network.neutron [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 992.485040] env[62692]: DEBUG nova.network.neutron [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Updating instance_info_cache with network_info: [{"id": "ab9d5209-bc8f-437d-a219-7f521cae10e6", "address": "fa:16:3e:d1:60:11", "network": {"id": "26fe2f12-b216-4fa5-a022-24c5ab315416", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-838584488-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb2ee16ec90c48e6babfb1d31ea3de94", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab9d5209-bc", "ovs_interfaceid": "ab9d5209-bc8f-437d-a219-7f521cae10e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.599834] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Releasing lock "refresh_cache-3f6c9744-a6e3-43f3-8b6b-624ba178747d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.695960] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]522eaf2c-ff18-a3fa-476f-9adedbb2b66c, 'name': SearchDatastore_Task, 'duration_secs': 0.00946} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.696258] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.696492] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 250b9cc2-706e-4fbf-887c-1c80fc27a0ed/250b9cc2-706e-4fbf-887c-1c80fc27a0ed.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 992.696748] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c2a12ef7-dae7-442f-a7ba-ca0f8e7bc70c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.703538] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 992.703538] env[62692]: value = "task-1141571" [ 992.703538] env[62692]: _type = "Task" [ 992.703538] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.711025] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141571, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.841806] env[62692]: DEBUG nova.compute.manager [req-40d5c779-a015-49c5-8d30-3287eb5d2984 req-42c7aba3-1d89-4725-8f38-bc69c18c1a36 service nova] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Received event network-vif-plugged-ab9d5209-bc8f-437d-a219-7f521cae10e6 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 992.842071] env[62692]: DEBUG oslo_concurrency.lockutils [req-40d5c779-a015-49c5-8d30-3287eb5d2984 req-42c7aba3-1d89-4725-8f38-bc69c18c1a36 service nova] Acquiring lock "7183ec4a-992d-48f1-8dda-7f499c2f4e1d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.842276] env[62692]: DEBUG oslo_concurrency.lockutils [req-40d5c779-a015-49c5-8d30-3287eb5d2984 req-42c7aba3-1d89-4725-8f38-bc69c18c1a36 service nova] Lock "7183ec4a-992d-48f1-8dda-7f499c2f4e1d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.842447] env[62692]: DEBUG oslo_concurrency.lockutils [req-40d5c779-a015-49c5-8d30-3287eb5d2984 req-42c7aba3-1d89-4725-8f38-bc69c18c1a36 service nova] Lock "7183ec4a-992d-48f1-8dda-7f499c2f4e1d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.842621] env[62692]: DEBUG nova.compute.manager [req-40d5c779-a015-49c5-8d30-3287eb5d2984 req-42c7aba3-1d89-4725-8f38-bc69c18c1a36 service nova] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] No waiting events found dispatching network-vif-plugged-ab9d5209-bc8f-437d-a219-7f521cae10e6 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 992.842799] env[62692]: WARNING nova.compute.manager [req-40d5c779-a015-49c5-8d30-3287eb5d2984 req-42c7aba3-1d89-4725-8f38-bc69c18c1a36 service nova] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Received unexpected event network-vif-plugged-ab9d5209-bc8f-437d-a219-7f521cae10e6 for instance with vm_state building and task_state spawning. [ 992.842961] env[62692]: DEBUG nova.compute.manager [req-40d5c779-a015-49c5-8d30-3287eb5d2984 req-42c7aba3-1d89-4725-8f38-bc69c18c1a36 service nova] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Received event network-changed-ab9d5209-bc8f-437d-a219-7f521cae10e6 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 992.843153] env[62692]: DEBUG nova.compute.manager [req-40d5c779-a015-49c5-8d30-3287eb5d2984 req-42c7aba3-1d89-4725-8f38-bc69c18c1a36 service nova] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Refreshing instance network info cache due to event network-changed-ab9d5209-bc8f-437d-a219-7f521cae10e6. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 992.843365] env[62692]: DEBUG oslo_concurrency.lockutils [req-40d5c779-a015-49c5-8d30-3287eb5d2984 req-42c7aba3-1d89-4725-8f38-bc69c18c1a36 service nova] Acquiring lock "refresh_cache-7183ec4a-992d-48f1-8dda-7f499c2f4e1d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.858501] env[62692]: DEBUG nova.scheduler.client.report [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 992.987642] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Releasing lock "refresh_cache-7183ec4a-992d-48f1-8dda-7f499c2f4e1d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.988158] env[62692]: DEBUG nova.compute.manager [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Instance network_info: |[{"id": "ab9d5209-bc8f-437d-a219-7f521cae10e6", "address": "fa:16:3e:d1:60:11", "network": {"id": "26fe2f12-b216-4fa5-a022-24c5ab315416", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-838584488-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb2ee16ec90c48e6babfb1d31ea3de94", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab9d5209-bc", "ovs_interfaceid": "ab9d5209-bc8f-437d-a219-7f521cae10e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 992.988536] env[62692]: DEBUG oslo_concurrency.lockutils [req-40d5c779-a015-49c5-8d30-3287eb5d2984 req-42c7aba3-1d89-4725-8f38-bc69c18c1a36 service nova] Acquired lock "refresh_cache-7183ec4a-992d-48f1-8dda-7f499c2f4e1d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.988775] env[62692]: DEBUG nova.network.neutron [req-40d5c779-a015-49c5-8d30-3287eb5d2984 req-42c7aba3-1d89-4725-8f38-bc69c18c1a36 service nova] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Refreshing network info cache for port ab9d5209-bc8f-437d-a219-7f521cae10e6 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 992.990366] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d1:60:11', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0636c3f6-fcb7-4954-ab07-c5cd0dee37b0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ab9d5209-bc8f-437d-a219-7f521cae10e6', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 992.999067] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Creating folder: Project (eb2ee16ec90c48e6babfb1d31ea3de94). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 993.001784] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-14f5d454-9a43-4da6-99b3-69b9ab2655bd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.011697] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Created folder: Project (eb2ee16ec90c48e6babfb1d31ea3de94) in parent group-v248868. [ 993.011978] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Creating folder: Instances. Parent ref: group-v249012. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 993.012144] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-41da4ee2-790b-464f-8d83-e5f3bbcb7043 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.023013] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Created folder: Instances in parent group-v249012. [ 993.023285] env[62692]: DEBUG oslo.service.loopingcall [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 993.023483] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 993.023695] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-52692ac8-e2e5-4e7c-9704-73588f55d450 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.044785] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 993.044785] env[62692]: value = "task-1141574" [ 993.044785] env[62692]: _type = "Task" [ 993.044785] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.052645] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141574, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.124061] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ebcabac-0d9b-4f9e-a5b9-2f2983e283cc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.145037] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac2f76d7-ad9a-4825-b771-eb125fc14527 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.154057] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Updating instance '3f6c9744-a6e3-43f3-8b6b-624ba178747d' progress to 83 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 993.213656] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141571, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.363366] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.372s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.363868] env[62692]: DEBUG nova.compute.manager [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 993.366767] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.700s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.366990] env[62692]: DEBUG nova.objects.instance [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lazy-loading 'resources' on Instance uuid 64a8329d-e0e9-4c2d-bd1f-844ee40eb980 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 993.557949] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141574, 'name': CreateVM_Task, 'duration_secs': 0.395288} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.558157] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 993.558802] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.558972] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.559487] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 993.559759] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-064808b9-6f44-4718-a263-4ab4ca674b9b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.565120] env[62692]: DEBUG oslo_vmware.api [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 993.565120] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52a40f24-1e18-e99c-9046-296715e7a163" [ 993.565120] env[62692]: _type = "Task" [ 993.565120] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.572607] env[62692]: DEBUG oslo_vmware.api [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52a40f24-1e18-e99c-9046-296715e7a163, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.661452] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 993.661801] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-11af4a35-e9c2-4b79-8436-62f487d18f46 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.669504] env[62692]: DEBUG oslo_vmware.api [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 993.669504] env[62692]: value = "task-1141575" [ 993.669504] env[62692]: _type = "Task" [ 993.669504] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.677393] env[62692]: DEBUG oslo_vmware.api [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141575, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.714613] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141571, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.519297} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.715308] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 250b9cc2-706e-4fbf-887c-1c80fc27a0ed/250b9cc2-706e-4fbf-887c-1c80fc27a0ed.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 993.715694] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 993.716089] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-27e58878-abaa-4fb6-af63-4276bf192a5d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.722748] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 993.722748] env[62692]: value = "task-1141576" [ 993.722748] env[62692]: _type = "Task" [ 993.722748] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.726670] env[62692]: DEBUG nova.network.neutron [req-40d5c779-a015-49c5-8d30-3287eb5d2984 req-42c7aba3-1d89-4725-8f38-bc69c18c1a36 service nova] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Updated VIF entry in instance network info cache for port ab9d5209-bc8f-437d-a219-7f521cae10e6. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 993.727355] env[62692]: DEBUG nova.network.neutron [req-40d5c779-a015-49c5-8d30-3287eb5d2984 req-42c7aba3-1d89-4725-8f38-bc69c18c1a36 service nova] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Updating instance_info_cache with network_info: [{"id": "ab9d5209-bc8f-437d-a219-7f521cae10e6", "address": "fa:16:3e:d1:60:11", "network": {"id": "26fe2f12-b216-4fa5-a022-24c5ab315416", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-838584488-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb2ee16ec90c48e6babfb1d31ea3de94", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab9d5209-bc", "ovs_interfaceid": "ab9d5209-bc8f-437d-a219-7f521cae10e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.736147] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141576, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.870189] env[62692]: DEBUG nova.compute.utils [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 993.874457] env[62692]: DEBUG nova.compute.manager [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 993.874457] env[62692]: DEBUG nova.network.neutron [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 993.914781] env[62692]: DEBUG nova.policy [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '249a3b566fcf4d88a973693512df370e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0b0b2e09a06a4fc7a00a03f72989ffed', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 994.075599] env[62692]: DEBUG oslo_vmware.api [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52a40f24-1e18-e99c-9046-296715e7a163, 'name': SearchDatastore_Task, 'duration_secs': 0.011747} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.075911] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.076166] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 994.076401] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.076547] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.076738] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 994.076978] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8e761bef-3be5-450a-9e3c-f51bf13b24f8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.095682] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 994.095905] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 994.096764] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e200c352-f90a-4283-9076-43edbcac9c1f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.102920] env[62692]: DEBUG oslo_vmware.api [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 994.102920] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52c1ac5d-d3a7-3318-6c46-b81e976ffd8b" [ 994.102920] env[62692]: _type = "Task" [ 994.102920] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.115410] env[62692]: DEBUG oslo_vmware.api [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52c1ac5d-d3a7-3318-6c46-b81e976ffd8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.179798] env[62692]: DEBUG oslo_vmware.api [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141575, 'name': PowerOnVM_Task, 'duration_secs': 0.468165} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.181984] env[62692]: DEBUG nova.network.neutron [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Successfully created port: dd324771-27d4-4505-b030-c8d8d9565968 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 994.183827] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 994.184074] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c2067f9a-2a7e-446b-9b07-c1ee02689261 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Updating instance '3f6c9744-a6e3-43f3-8b6b-624ba178747d' progress to 100 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 994.217568] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-053e997d-8fe3-45c4-9d1c-d1ced7d8ef8e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.228374] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-695d097e-bdaf-4dbb-ac00-fcc36af50723 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.232218] env[62692]: DEBUG oslo_concurrency.lockutils [req-40d5c779-a015-49c5-8d30-3287eb5d2984 req-42c7aba3-1d89-4725-8f38-bc69c18c1a36 service nova] Releasing lock "refresh_cache-7183ec4a-992d-48f1-8dda-7f499c2f4e1d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.237968] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141576, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077091} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.263369] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 994.264737] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de4ec56-63b4-4d81-9246-7d1a639eec8f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.267714] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1590385-55b8-484b-b0b6-2e44a7c9af91 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.284228] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3afcf621-bf2f-4062-8ef5-22ff8c274567 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.296554] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] 250b9cc2-706e-4fbf-887c-1c80fc27a0ed/250b9cc2-706e-4fbf-887c-1c80fc27a0ed.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 994.297094] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a4777635-75da-4e24-a398-1c7c06d3258e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.321186] env[62692]: DEBUG nova.compute.provider_tree [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 994.324032] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 994.324032] env[62692]: value = "task-1141577" [ 994.324032] env[62692]: _type = "Task" [ 994.324032] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.332710] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141577, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.377235] env[62692]: DEBUG nova.compute.manager [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 994.614132] env[62692]: DEBUG oslo_vmware.api [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52c1ac5d-d3a7-3318-6c46-b81e976ffd8b, 'name': SearchDatastore_Task, 'duration_secs': 0.028529} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.615117] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1cd587a9-5175-41e7-b444-7acf45ed6426 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.621834] env[62692]: DEBUG oslo_vmware.api [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 994.621834] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52b1b8fe-6b1d-adc2-d2b7-09c980993c90" [ 994.621834] env[62692]: _type = "Task" [ 994.621834] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.630897] env[62692]: DEBUG oslo_vmware.api [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52b1b8fe-6b1d-adc2-d2b7-09c980993c90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.883574] env[62692]: DEBUG nova.scheduler.client.report [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 994.883574] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141577, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.134933] env[62692]: DEBUG oslo_vmware.api [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52b1b8fe-6b1d-adc2-d2b7-09c980993c90, 'name': SearchDatastore_Task, 'duration_secs': 0.010652} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.135200] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.135461] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 7183ec4a-992d-48f1-8dda-7f499c2f4e1d/7183ec4a-992d-48f1-8dda-7f499c2f4e1d.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 995.135725] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-66bfe968-dffb-47fd-8b05-d1de758915ce {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.142106] env[62692]: DEBUG oslo_vmware.api [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 995.142106] env[62692]: value = "task-1141578" [ 995.142106] env[62692]: _type = "Task" [ 995.142106] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.149956] env[62692]: DEBUG oslo_vmware.api [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141578, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.339906] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141577, 'name': ReconfigVM_Task, 'duration_secs': 0.578729} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.340250] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Reconfigured VM instance instance-00000054 to attach disk [datastore1] 250b9cc2-706e-4fbf-887c-1c80fc27a0ed/250b9cc2-706e-4fbf-887c-1c80fc27a0ed.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 995.341409] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.975s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.343252] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a84172a4-1de9-468a-8c1e-56cdd2fa8509 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.345103] env[62692]: DEBUG oslo_concurrency.lockutils [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.291s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.345332] env[62692]: DEBUG nova.objects.instance [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lazy-loading 'resources' on Instance uuid 46bce865-188b-487c-a73f-cf524059c797 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 995.351029] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 995.351029] env[62692]: value = "task-1141579" [ 995.351029] env[62692]: _type = "Task" [ 995.351029] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.359882] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141579, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.364387] env[62692]: INFO nova.scheduler.client.report [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Deleted allocations for instance 64a8329d-e0e9-4c2d-bd1f-844ee40eb980 [ 995.386110] env[62692]: DEBUG nova.compute.manager [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 995.411846] env[62692]: DEBUG nova.virt.hardware [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 995.412915] env[62692]: DEBUG nova.virt.hardware [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 995.412915] env[62692]: DEBUG nova.virt.hardware [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 995.412915] env[62692]: DEBUG nova.virt.hardware [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 995.412915] env[62692]: DEBUG nova.virt.hardware [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 995.413142] env[62692]: DEBUG nova.virt.hardware [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 995.413236] env[62692]: DEBUG nova.virt.hardware [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 995.413407] env[62692]: DEBUG nova.virt.hardware [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 995.413581] env[62692]: DEBUG nova.virt.hardware [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 995.414405] env[62692]: DEBUG nova.virt.hardware [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 995.414405] env[62692]: DEBUG nova.virt.hardware [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 995.415071] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e986571-eb6d-46d9-b42a-551f55ff9a28 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.423526] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24be34a1-95ce-40e9-8c45-9054ab25377d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.652711] env[62692]: DEBUG oslo_vmware.api [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141578, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.661194] env[62692]: DEBUG nova.compute.manager [req-7e0344d8-b8b9-4f30-a699-a4fcf5ff041e req-13e1a151-8228-4ee6-9c72-e5a67d348b92 service nova] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Received event network-vif-plugged-dd324771-27d4-4505-b030-c8d8d9565968 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 995.661372] env[62692]: DEBUG oslo_concurrency.lockutils [req-7e0344d8-b8b9-4f30-a699-a4fcf5ff041e req-13e1a151-8228-4ee6-9c72-e5a67d348b92 service nova] Acquiring lock "f6085887-a5d3-4096-be38-d843bb54304a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.661584] env[62692]: DEBUG oslo_concurrency.lockutils [req-7e0344d8-b8b9-4f30-a699-a4fcf5ff041e req-13e1a151-8228-4ee6-9c72-e5a67d348b92 service nova] Lock "f6085887-a5d3-4096-be38-d843bb54304a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.661976] env[62692]: DEBUG oslo_concurrency.lockutils [req-7e0344d8-b8b9-4f30-a699-a4fcf5ff041e req-13e1a151-8228-4ee6-9c72-e5a67d348b92 service nova] Lock "f6085887-a5d3-4096-be38-d843bb54304a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.662207] env[62692]: DEBUG nova.compute.manager [req-7e0344d8-b8b9-4f30-a699-a4fcf5ff041e req-13e1a151-8228-4ee6-9c72-e5a67d348b92 service nova] [instance: f6085887-a5d3-4096-be38-d843bb54304a] No waiting events found dispatching network-vif-plugged-dd324771-27d4-4505-b030-c8d8d9565968 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 995.662387] env[62692]: WARNING nova.compute.manager [req-7e0344d8-b8b9-4f30-a699-a4fcf5ff041e req-13e1a151-8228-4ee6-9c72-e5a67d348b92 service nova] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Received unexpected event network-vif-plugged-dd324771-27d4-4505-b030-c8d8d9565968 for instance with vm_state building and task_state spawning. [ 995.759113] env[62692]: DEBUG nova.network.neutron [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Successfully updated port: dd324771-27d4-4505-b030-c8d8d9565968 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 995.861159] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141579, 'name': Rename_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.872641] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c0c48d26-ea3e-493d-ad23-db1ac8b940b5 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "64a8329d-e0e9-4c2d-bd1f-844ee40eb980" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.396s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.132249] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-227a9a68-bb22-47da-9a49-52ab7b259ad2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.139649] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d1d0827-7c5c-416f-b224-42d634e76570 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.152754] env[62692]: DEBUG oslo_vmware.api [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141578, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.177289] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22ad187a-f09e-40b6-9126-8d102301979e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.187763] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62bdf5db-18db-45d9-9dc5-69d0c2a14dcc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.206049] env[62692]: DEBUG nova.compute.provider_tree [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 996.261809] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "refresh_cache-f6085887-a5d3-4096-be38-d843bb54304a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.261809] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquired lock "refresh_cache-f6085887-a5d3-4096-be38-d843bb54304a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.261992] env[62692]: DEBUG nova.network.neutron [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 996.362300] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141579, 'name': Rename_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.654292] env[62692]: DEBUG oslo_vmware.api [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141578, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.376397} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.654587] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 7183ec4a-992d-48f1-8dda-7f499c2f4e1d/7183ec4a-992d-48f1-8dda-7f499c2f4e1d.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 996.654812] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 996.655077] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6a31f3b8-c24d-4404-b427-5f8b3641c2f8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.662420] env[62692]: DEBUG oslo_vmware.api [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 996.662420] env[62692]: value = "task-1141580" [ 996.662420] env[62692]: _type = "Task" [ 996.662420] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.672228] env[62692]: DEBUG oslo_vmware.api [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141580, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.707997] env[62692]: DEBUG nova.scheduler.client.report [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 996.793346] env[62692]: DEBUG nova.network.neutron [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 996.838906] env[62692]: DEBUG oslo_concurrency.lockutils [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "3f6c9744-a6e3-43f3-8b6b-624ba178747d" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.839208] env[62692]: DEBUG oslo_concurrency.lockutils [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "3f6c9744-a6e3-43f3-8b6b-624ba178747d" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.839419] env[62692]: DEBUG nova.compute.manager [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Going to confirm migration 2 {{(pid=62692) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 996.864831] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141579, 'name': Rename_Task, 'duration_secs': 1.288401} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.865143] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 996.865399] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0f2ffb2b-7023-49d8-9f17-bc326d67810a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.872497] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 996.872497] env[62692]: value = "task-1141581" [ 996.872497] env[62692]: _type = "Task" [ 996.872497] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.881530] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141581, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.940562] env[62692]: DEBUG nova.network.neutron [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Updating instance_info_cache with network_info: [{"id": "dd324771-27d4-4505-b030-c8d8d9565968", "address": "fa:16:3e:a7:80:84", "network": {"id": "2c82845f-101b-48c4-bd0b-4704573a3615", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1972972124-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b0b2e09a06a4fc7a00a03f72989ffed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16f09e8c-5240-4839-80cc-62ec29700bd2", "external-id": "nsx-vlan-transportzone-720", "segmentation_id": 720, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd324771-27", "ovs_interfaceid": "dd324771-27d4-4505-b030-c8d8d9565968", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.172456] env[62692]: DEBUG oslo_vmware.api [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141580, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070139} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.172753] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 997.173516] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6970a760-b92b-4695-9980-626d19942e86 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.196099] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] 7183ec4a-992d-48f1-8dda-7f499c2f4e1d/7183ec4a-992d-48f1-8dda-7f499c2f4e1d.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 997.196099] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b191b676-8611-4a07-9195-f9419703140f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.213652] env[62692]: DEBUG oslo_concurrency.lockutils [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.868s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.216585] env[62692]: DEBUG oslo_concurrency.lockutils [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.109s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.216585] env[62692]: DEBUG nova.objects.instance [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lazy-loading 'resources' on Instance uuid a668a569-a90e-4e59-8d2e-b0225745b500 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.220227] env[62692]: DEBUG oslo_vmware.api [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 997.220227] env[62692]: value = "task-1141582" [ 997.220227] env[62692]: _type = "Task" [ 997.220227] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.227806] env[62692]: DEBUG oslo_vmware.api [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141582, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.235567] env[62692]: INFO nova.scheduler.client.report [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Deleted allocations for instance 46bce865-188b-487c-a73f-cf524059c797 [ 997.382422] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141581, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.413258] env[62692]: DEBUG oslo_concurrency.lockutils [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "refresh_cache-3f6c9744-a6e3-43f3-8b6b-624ba178747d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.413500] env[62692]: DEBUG oslo_concurrency.lockutils [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired lock "refresh_cache-3f6c9744-a6e3-43f3-8b6b-624ba178747d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.413724] env[62692]: DEBUG nova.network.neutron [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 997.413950] env[62692]: DEBUG nova.objects.instance [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lazy-loading 'info_cache' on Instance uuid 3f6c9744-a6e3-43f3-8b6b-624ba178747d {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.446203] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Releasing lock "refresh_cache-f6085887-a5d3-4096-be38-d843bb54304a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.446203] env[62692]: DEBUG nova.compute.manager [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Instance network_info: |[{"id": "dd324771-27d4-4505-b030-c8d8d9565968", "address": "fa:16:3e:a7:80:84", "network": {"id": "2c82845f-101b-48c4-bd0b-4704573a3615", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1972972124-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b0b2e09a06a4fc7a00a03f72989ffed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16f09e8c-5240-4839-80cc-62ec29700bd2", "external-id": "nsx-vlan-transportzone-720", "segmentation_id": 720, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd324771-27", "ovs_interfaceid": "dd324771-27d4-4505-b030-c8d8d9565968", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 997.446203] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a7:80:84', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '16f09e8c-5240-4839-80cc-62ec29700bd2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dd324771-27d4-4505-b030-c8d8d9565968', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 997.452151] env[62692]: DEBUG oslo.service.loopingcall [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 997.452408] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 997.452676] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e8f7e9e4-a6a0-4ad4-b54b-ffc6141300b2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.474181] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 997.474181] env[62692]: value = "task-1141583" [ 997.474181] env[62692]: _type = "Task" [ 997.474181] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.482792] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141583, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.730574] env[62692]: DEBUG nova.compute.manager [req-9eb7e275-e4be-43d4-ada7-2015c962d4fa req-55f47cd4-19c2-4c8d-bf78-a3091f1b5d0b service nova] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Received event network-changed-dd324771-27d4-4505-b030-c8d8d9565968 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 997.730892] env[62692]: DEBUG nova.compute.manager [req-9eb7e275-e4be-43d4-ada7-2015c962d4fa req-55f47cd4-19c2-4c8d-bf78-a3091f1b5d0b service nova] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Refreshing instance network info cache due to event network-changed-dd324771-27d4-4505-b030-c8d8d9565968. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 997.731204] env[62692]: DEBUG oslo_concurrency.lockutils [req-9eb7e275-e4be-43d4-ada7-2015c962d4fa req-55f47cd4-19c2-4c8d-bf78-a3091f1b5d0b service nova] Acquiring lock "refresh_cache-f6085887-a5d3-4096-be38-d843bb54304a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.731423] env[62692]: DEBUG oslo_concurrency.lockutils [req-9eb7e275-e4be-43d4-ada7-2015c962d4fa req-55f47cd4-19c2-4c8d-bf78-a3091f1b5d0b service nova] Acquired lock "refresh_cache-f6085887-a5d3-4096-be38-d843bb54304a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.731673] env[62692]: DEBUG nova.network.neutron [req-9eb7e275-e4be-43d4-ada7-2015c962d4fa req-55f47cd4-19c2-4c8d-bf78-a3091f1b5d0b service nova] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Refreshing network info cache for port dd324771-27d4-4505-b030-c8d8d9565968 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 997.743179] env[62692]: DEBUG oslo_vmware.api [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141582, 'name': ReconfigVM_Task, 'duration_secs': 0.348482} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.743680] env[62692]: DEBUG oslo_concurrency.lockutils [None req-54fc3e54-c472-47a4-a780-57a1afb23508 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "46bce865-188b-487c-a73f-cf524059c797" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.837s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.744637] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Reconfigured VM instance instance-00000055 to attach disk [datastore1] 7183ec4a-992d-48f1-8dda-7f499c2f4e1d/7183ec4a-992d-48f1-8dda-7f499c2f4e1d.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 997.745662] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9d96af61-3d57-4edc-b3ba-e3fb77042c01 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.753052] env[62692]: DEBUG oslo_vmware.api [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 997.753052] env[62692]: value = "task-1141584" [ 997.753052] env[62692]: _type = "Task" [ 997.753052] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.763558] env[62692]: DEBUG oslo_vmware.api [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141584, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.885410] env[62692]: DEBUG oslo_vmware.api [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141581, 'name': PowerOnVM_Task, 'duration_secs': 0.679256} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.885691] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 997.885893] env[62692]: INFO nova.compute.manager [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Took 9.34 seconds to spawn the instance on the hypervisor. [ 997.886089] env[62692]: DEBUG nova.compute.manager [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 997.886892] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbc5dc30-5520-42ce-b2a0-bb1a54d95233 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.985716] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141583, 'name': CreateVM_Task, 'duration_secs': 0.370098} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.985915] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 997.986609] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.986949] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.987100] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 997.987688] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a864a44f-4b45-43b5-a33b-1bd1a3300e4c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.991916] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 997.991916] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]527814a2-53ba-2415-31a2-6edd4b741de4" [ 997.991916] env[62692]: _type = "Task" [ 997.991916] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.003339] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]527814a2-53ba-2415-31a2-6edd4b741de4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.034559] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84c7c7f9-ceb2-4ec6-933d-6d430919a4de {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.041444] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94db4db5-fdcf-423d-b2cf-7ee5083471eb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.072378] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d97d1ecf-cacf-4f06-9ad4-9ebf97e91148 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.080276] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e1dd5cf-6917-4e57-b5e0-c8a90eb269a4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.093222] env[62692]: DEBUG nova.compute.provider_tree [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 998.264497] env[62692]: DEBUG oslo_vmware.api [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141584, 'name': Rename_Task, 'duration_secs': 0.149357} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.264882] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 998.265033] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3c1269cc-1bdd-4032-a2ab-2136e4e83b47 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.273111] env[62692]: DEBUG oslo_vmware.api [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 998.273111] env[62692]: value = "task-1141585" [ 998.273111] env[62692]: _type = "Task" [ 998.273111] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.280814] env[62692]: DEBUG oslo_vmware.api [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141585, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.414593] env[62692]: INFO nova.compute.manager [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Took 31.43 seconds to build instance. [ 998.505046] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]527814a2-53ba-2415-31a2-6edd4b741de4, 'name': SearchDatastore_Task, 'duration_secs': 0.024019} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.506910] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.507273] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 998.507479] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.507645] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.507821] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 998.508115] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2204c6d3-8520-4b18-86fe-8394075e631f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.517284] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 998.517535] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 998.518333] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99d80838-6bfc-4f52-b731-80949e081649 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.523484] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 998.523484] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52206c34-d04e-c323-7a1b-c266e1168290" [ 998.523484] env[62692]: _type = "Task" [ 998.523484] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.527394] env[62692]: DEBUG nova.network.neutron [req-9eb7e275-e4be-43d4-ada7-2015c962d4fa req-55f47cd4-19c2-4c8d-bf78-a3091f1b5d0b service nova] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Updated VIF entry in instance network info cache for port dd324771-27d4-4505-b030-c8d8d9565968. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 998.527746] env[62692]: DEBUG nova.network.neutron [req-9eb7e275-e4be-43d4-ada7-2015c962d4fa req-55f47cd4-19c2-4c8d-bf78-a3091f1b5d0b service nova] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Updating instance_info_cache with network_info: [{"id": "dd324771-27d4-4505-b030-c8d8d9565968", "address": "fa:16:3e:a7:80:84", "network": {"id": "2c82845f-101b-48c4-bd0b-4704573a3615", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1972972124-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b0b2e09a06a4fc7a00a03f72989ffed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16f09e8c-5240-4839-80cc-62ec29700bd2", "external-id": "nsx-vlan-transportzone-720", "segmentation_id": 720, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd324771-27", "ovs_interfaceid": "dd324771-27d4-4505-b030-c8d8d9565968", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.534815] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52206c34-d04e-c323-7a1b-c266e1168290, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.596773] env[62692]: DEBUG nova.scheduler.client.report [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 998.782962] env[62692]: DEBUG oslo_vmware.api [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141585, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.918373] env[62692]: DEBUG oslo_concurrency.lockutils [None req-01d0d9ee-3565-4735-abb2-fdfab76eae53 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "250b9cc2-706e-4fbf-887c-1c80fc27a0ed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.952s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.974111] env[62692]: DEBUG nova.network.neutron [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Updating instance_info_cache with network_info: [{"id": "2cb4b52c-0d89-4402-ae48-dc6f226c557c", "address": "fa:16:3e:c8:b1:41", "network": {"id": "f6352999-f943-431d-b54a-c9dc1d810a85", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-121037111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9cb8ae08824245a7911de7455e7fcb68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c842425c-544e-4ce2-9657-512723bd318e", "external-id": "nsx-vlan-transportzone-80", "segmentation_id": 80, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cb4b52c-0d", "ovs_interfaceid": "2cb4b52c-0d89-4402-ae48-dc6f226c557c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.033795] env[62692]: DEBUG oslo_concurrency.lockutils [req-9eb7e275-e4be-43d4-ada7-2015c962d4fa req-55f47cd4-19c2-4c8d-bf78-a3091f1b5d0b service nova] Releasing lock "refresh_cache-f6085887-a5d3-4096-be38-d843bb54304a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.034278] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52206c34-d04e-c323-7a1b-c266e1168290, 'name': SearchDatastore_Task, 'duration_secs': 0.014321} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.035032] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ba27235-baba-4d57-bd54-e48056627c48 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.040250] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 999.040250] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5284c560-77c4-fa6b-80f9-88633cdb293f" [ 999.040250] env[62692]: _type = "Task" [ 999.040250] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.048289] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5284c560-77c4-fa6b-80f9-88633cdb293f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.101842] env[62692]: DEBUG oslo_concurrency.lockutils [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.886s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.104354] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.378s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.105365] env[62692]: DEBUG nova.objects.instance [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Lazy-loading 'resources' on Instance uuid fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.122077] env[62692]: INFO nova.scheduler.client.report [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Deleted allocations for instance a668a569-a90e-4e59-8d2e-b0225745b500 [ 999.283801] env[62692]: DEBUG oslo_vmware.api [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141585, 'name': PowerOnVM_Task, 'duration_secs': 0.655227} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.284122] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 999.284322] env[62692]: INFO nova.compute.manager [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Took 8.30 seconds to spawn the instance on the hypervisor. [ 999.284552] env[62692]: DEBUG nova.compute.manager [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 999.285385] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daa5a558-f720-4eae-8b80-d3188be10b1e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.476997] env[62692]: DEBUG oslo_concurrency.lockutils [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Releasing lock "refresh_cache-3f6c9744-a6e3-43f3-8b6b-624ba178747d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.477426] env[62692]: DEBUG nova.objects.instance [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lazy-loading 'migration_context' on Instance uuid 3f6c9744-a6e3-43f3-8b6b-624ba178747d {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.550695] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5284c560-77c4-fa6b-80f9-88633cdb293f, 'name': SearchDatastore_Task, 'duration_secs': 0.012254} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.550971] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.551255] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] f6085887-a5d3-4096-be38-d843bb54304a/f6085887-a5d3-4096-be38-d843bb54304a.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 999.551523] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ddfc6d57-26e6-4228-85b9-138484950d7e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.558314] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 999.558314] env[62692]: value = "task-1141586" [ 999.558314] env[62692]: _type = "Task" [ 999.558314] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.567031] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141586, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.632168] env[62692]: DEBUG oslo_concurrency.lockutils [None req-60ed5233-bbbb-44b1-83cd-956af0a0cc52 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "a668a569-a90e-4e59-8d2e-b0225745b500" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.634s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.808560] env[62692]: INFO nova.compute.manager [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Took 31.68 seconds to build instance. [ 999.897839] env[62692]: DEBUG oslo_concurrency.lockutils [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "250b9cc2-706e-4fbf-887c-1c80fc27a0ed" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.898265] env[62692]: DEBUG oslo_concurrency.lockutils [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "250b9cc2-706e-4fbf-887c-1c80fc27a0ed" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.898532] env[62692]: INFO nova.compute.manager [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Shelving [ 999.980847] env[62692]: DEBUG nova.compute.manager [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Stashing vm_state: active {{(pid=62692) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 999.984744] env[62692]: DEBUG nova.objects.base [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Object Instance<3f6c9744-a6e3-43f3-8b6b-624ba178747d> lazy-loaded attributes: info_cache,migration_context {{(pid=62692) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 999.989305] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c36a353c-b106-4eaf-bb4c-fea5bb20cfce {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.012281] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0abd0cc0-253e-4bbe-b2e6-2938f89cba87 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.023806] env[62692]: DEBUG oslo_vmware.api [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1000.023806] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52d48dfd-9ce5-1718-9bbe-156c7e5244ba" [ 1000.023806] env[62692]: _type = "Task" [ 1000.023806] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.034149] env[62692]: DEBUG oslo_vmware.api [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52d48dfd-9ce5-1718-9bbe-156c7e5244ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.040101] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8ce7d85-8154-410a-872d-79cd2a9e65f8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.047958] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b93ae45c-0fe4-4dd5-83c3-b0ef3dd113f3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.084490] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1458cbe-cc0a-417b-93ac-990e676354b8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.096109] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc98a8a7-9687-41f1-a31e-f32748ed71b3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.102054] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141586, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.110865] env[62692]: DEBUG nova.compute.provider_tree [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1000.285189] env[62692]: DEBUG nova.compute.manager [req-e12010e5-b084-4c21-ae2e-96aef0645877 req-489ec97f-7636-49af-8665-87e865a525be service nova] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Received event network-changed-ab9d5209-bc8f-437d-a219-7f521cae10e6 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1000.285432] env[62692]: DEBUG nova.compute.manager [req-e12010e5-b084-4c21-ae2e-96aef0645877 req-489ec97f-7636-49af-8665-87e865a525be service nova] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Refreshing instance network info cache due to event network-changed-ab9d5209-bc8f-437d-a219-7f521cae10e6. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1000.285719] env[62692]: DEBUG oslo_concurrency.lockutils [req-e12010e5-b084-4c21-ae2e-96aef0645877 req-489ec97f-7636-49af-8665-87e865a525be service nova] Acquiring lock "refresh_cache-7183ec4a-992d-48f1-8dda-7f499c2f4e1d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.285924] env[62692]: DEBUG oslo_concurrency.lockutils [req-e12010e5-b084-4c21-ae2e-96aef0645877 req-489ec97f-7636-49af-8665-87e865a525be service nova] Acquired lock "refresh_cache-7183ec4a-992d-48f1-8dda-7f499c2f4e1d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.286161] env[62692]: DEBUG nova.network.neutron [req-e12010e5-b084-4c21-ae2e-96aef0645877 req-489ec97f-7636-49af-8665-87e865a525be service nova] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Refreshing network info cache for port ab9d5209-bc8f-437d-a219-7f521cae10e6 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1000.311645] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d19831e7-6b05-45d2-9b6c-153a9e481dc0 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "7183ec4a-992d-48f1-8dda-7f499c2f4e1d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.195s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.407380] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1000.407669] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba821f7a-1cd2-45a7-a2f7-30706b63be8c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.415717] env[62692]: DEBUG oslo_vmware.api [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1000.415717] env[62692]: value = "task-1141587" [ 1000.415717] env[62692]: _type = "Task" [ 1000.415717] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.425153] env[62692]: DEBUG oslo_vmware.api [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141587, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.503248] env[62692]: DEBUG oslo_concurrency.lockutils [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.535359] env[62692]: DEBUG oslo_vmware.api [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52d48dfd-9ce5-1718-9bbe-156c7e5244ba, 'name': SearchDatastore_Task, 'duration_secs': 0.028412} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.535887] env[62692]: DEBUG oslo_concurrency.lockutils [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.592425] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141586, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.575227} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.592908] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] f6085887-a5d3-4096-be38-d843bb54304a/f6085887-a5d3-4096-be38-d843bb54304a.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1000.593091] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1000.593375] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2c927513-7085-4c26-bfe4-040df5b3ad9e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.602073] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 1000.602073] env[62692]: value = "task-1141588" [ 1000.602073] env[62692]: _type = "Task" [ 1000.602073] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.614399] env[62692]: DEBUG nova.scheduler.client.report [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1000.618526] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141588, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.925328] env[62692]: DEBUG oslo_vmware.api [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141587, 'name': PowerOffVM_Task, 'duration_secs': 0.361175} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.925624] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1000.926438] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b2bdd8f-0f38-4d48-a0c9-e5dff05b31a4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.953705] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ae603dc-8b44-4deb-a1db-9e36e493dbb7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.111543] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141588, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091979} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.111797] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1001.112571] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-902a6cb0-b276-4431-8d5f-0c4f22a583e9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.126646] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.022s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.138186] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] f6085887-a5d3-4096-be38-d843bb54304a/f6085887-a5d3-4096-be38-d843bb54304a.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1001.139457] env[62692]: DEBUG nova.network.neutron [req-e12010e5-b084-4c21-ae2e-96aef0645877 req-489ec97f-7636-49af-8665-87e865a525be service nova] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Updated VIF entry in instance network info cache for port ab9d5209-bc8f-437d-a219-7f521cae10e6. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1001.139637] env[62692]: DEBUG nova.network.neutron [req-e12010e5-b084-4c21-ae2e-96aef0645877 req-489ec97f-7636-49af-8665-87e865a525be service nova] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Updating instance_info_cache with network_info: [{"id": "ab9d5209-bc8f-437d-a219-7f521cae10e6", "address": "fa:16:3e:d1:60:11", "network": {"id": "26fe2f12-b216-4fa5-a022-24c5ab315416", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-838584488-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb2ee16ec90c48e6babfb1d31ea3de94", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab9d5209-bc", "ovs_interfaceid": "ab9d5209-bc8f-437d-a219-7f521cae10e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.140820] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.849s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.142269] env[62692]: INFO nova.compute.claims [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1001.146780] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7bea2c4d-98d7-4f32-b5f2-fedcfe3a490b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.161803] env[62692]: INFO nova.scheduler.client.report [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Deleted allocations for instance fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957 [ 1001.171038] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 1001.171038] env[62692]: value = "task-1141589" [ 1001.171038] env[62692]: _type = "Task" [ 1001.171038] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.181445] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141589, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.465714] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Creating Snapshot of the VM instance {{(pid=62692) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1001.466193] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-ff70037b-bc2d-451c-801c-2f2fc6a60144 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.480551] env[62692]: DEBUG oslo_vmware.api [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1001.480551] env[62692]: value = "task-1141590" [ 1001.480551] env[62692]: _type = "Task" [ 1001.480551] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.489631] env[62692]: DEBUG oslo_vmware.api [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141590, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.663708] env[62692]: DEBUG oslo_concurrency.lockutils [req-e12010e5-b084-4c21-ae2e-96aef0645877 req-489ec97f-7636-49af-8665-87e865a525be service nova] Releasing lock "refresh_cache-7183ec4a-992d-48f1-8dda-7f499c2f4e1d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.672301] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f201ff31-5711-4d04-abf8-b29db47d5eb5 tempest-ServerShowV247Test-1998851077 tempest-ServerShowV247Test-1998851077-project-member] Lock "fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.292s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.684487] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141589, 'name': ReconfigVM_Task, 'duration_secs': 0.316954} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.684851] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Reconfigured VM instance instance-00000056 to attach disk [datastore1] f6085887-a5d3-4096-be38-d843bb54304a/f6085887-a5d3-4096-be38-d843bb54304a.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1001.685667] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0f56fb22-661f-478b-87ca-99c6fde2416a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.693137] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 1001.693137] env[62692]: value = "task-1141591" [ 1001.693137] env[62692]: _type = "Task" [ 1001.693137] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.708911] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141591, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.990795] env[62692]: DEBUG oslo_vmware.api [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141590, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.205317] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141591, 'name': Rename_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.315247] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "e8f645da-3b4a-4f21-9a17-aebdc8d112df" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.315504] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "e8f645da-3b4a-4f21-9a17-aebdc8d112df" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.471481] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e51986e5-e0c1-4d11-8f1d-d74d5cea0529 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.479044] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dd2e919-6b0c-449d-bbf3-9edcba191fcc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.489962] env[62692]: DEBUG oslo_vmware.api [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141590, 'name': CreateSnapshot_Task, 'duration_secs': 0.602505} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.516739] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Created Snapshot of the VM instance {{(pid=62692) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1002.517273] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e8a45ff-58d0-436f-971b-c8c83c0c5f04 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.520856] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff0e0d78-21cb-456f-9839-eed0ffbbcb85 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.531478] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c137d45-4b45-4b3d-8575-e0352822e4c1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.547042] env[62692]: DEBUG nova.compute.provider_tree [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.704668] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141591, 'name': Rename_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.819679] env[62692]: DEBUG nova.compute.manager [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1003.045463] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Creating linked-clone VM from snapshot {{(pid=62692) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1003.045822] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-8425e91c-ce28-45e3-ac55-3ea6510601b2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.049600] env[62692]: DEBUG nova.scheduler.client.report [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1003.059331] env[62692]: DEBUG oslo_vmware.api [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1003.059331] env[62692]: value = "task-1141592" [ 1003.059331] env[62692]: _type = "Task" [ 1003.059331] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.068728] env[62692]: DEBUG oslo_vmware.api [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141592, 'name': CloneVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.206257] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141591, 'name': Rename_Task, 'duration_secs': 1.151219} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.206480] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1003.206718] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cf41300d-aafa-4c1c-bcc3-5c9af1283330 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.213559] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 1003.213559] env[62692]: value = "task-1141593" [ 1003.213559] env[62692]: _type = "Task" [ 1003.213559] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.221401] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141593, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.342516] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.554473] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.413s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.555165] env[62692]: DEBUG nova.compute.manager [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1003.558162] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.258s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.558439] env[62692]: DEBUG nova.objects.instance [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Lazy-loading 'resources' on Instance uuid 084ce29c-302a-45cc-b36f-0143eccee298 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1003.571146] env[62692]: DEBUG oslo_vmware.api [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141592, 'name': CloneVM_Task} progress is 94%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.725833] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141593, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.063841] env[62692]: DEBUG nova.compute.utils [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1004.065991] env[62692]: DEBUG nova.compute.manager [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1004.065991] env[62692]: DEBUG nova.network.neutron [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1004.077237] env[62692]: DEBUG oslo_vmware.api [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141592, 'name': CloneVM_Task} progress is 95%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.135107] env[62692]: DEBUG nova.policy [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f9b3e1975b8444889bd4a6fd947367a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7aeb525d0d3644b5805b26a1a12fb450', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1004.226327] env[62692]: DEBUG oslo_vmware.api [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141593, 'name': PowerOnVM_Task, 'duration_secs': 0.554393} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.226612] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1004.226818] env[62692]: INFO nova.compute.manager [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Took 8.84 seconds to spawn the instance on the hypervisor. [ 1004.227015] env[62692]: DEBUG nova.compute.manager [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1004.227787] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ff62ea5-13d3-499c-9f6e-f07faf691b4a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.384320] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9b69654-fd1e-4091-8bd0-1bdc322d0a60 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.392340] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1193120d-fe58-4cf7-87e0-1806cba58d39 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.423678] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faa7c20c-ebb2-4ae9-90e8-7d18710067c1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.431558] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c49fd4a3-1744-4aa8-8422-0cf2a8e78322 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.445721] env[62692]: DEBUG nova.compute.provider_tree [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1004.505414] env[62692]: DEBUG nova.network.neutron [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Successfully created port: 9867c68f-a0ff-499d-9b91-47cbd25c4b69 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1004.574730] env[62692]: DEBUG nova.compute.manager [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1004.577315] env[62692]: DEBUG oslo_vmware.api [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141592, 'name': CloneVM_Task, 'duration_secs': 1.311811} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.577815] env[62692]: INFO nova.virt.vmwareapi.vmops [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Created linked-clone VM from snapshot [ 1004.581017] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd481204-1898-4306-91ef-b90b9e6107f7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.586061] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Uploading image 01209948-5245-453f-bb56-fb80a8ea2d67 {{(pid=62692) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1004.614189] env[62692]: DEBUG oslo_vmware.rw_handles [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1004.614189] env[62692]: value = "vm-249017" [ 1004.614189] env[62692]: _type = "VirtualMachine" [ 1004.614189] env[62692]: }. {{(pid=62692) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1004.614747] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-ddc65111-cd6d-415c-a514-f772455df28a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.622194] env[62692]: DEBUG oslo_vmware.rw_handles [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lease: (returnval){ [ 1004.622194] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]526320df-b908-c34f-34ab-a4077bdec8a0" [ 1004.622194] env[62692]: _type = "HttpNfcLease" [ 1004.622194] env[62692]: } obtained for exporting VM: (result){ [ 1004.622194] env[62692]: value = "vm-249017" [ 1004.622194] env[62692]: _type = "VirtualMachine" [ 1004.622194] env[62692]: }. {{(pid=62692) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1004.622573] env[62692]: DEBUG oslo_vmware.api [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the lease: (returnval){ [ 1004.622573] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]526320df-b908-c34f-34ab-a4077bdec8a0" [ 1004.622573] env[62692]: _type = "HttpNfcLease" [ 1004.622573] env[62692]: } to be ready. {{(pid=62692) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1004.629095] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1004.629095] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]526320df-b908-c34f-34ab-a4077bdec8a0" [ 1004.629095] env[62692]: _type = "HttpNfcLease" [ 1004.629095] env[62692]: } is initializing. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1004.746756] env[62692]: INFO nova.compute.manager [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Took 34.43 seconds to build instance. [ 1004.948628] env[62692]: DEBUG nova.scheduler.client.report [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1005.131955] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1005.131955] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]526320df-b908-c34f-34ab-a4077bdec8a0" [ 1005.131955] env[62692]: _type = "HttpNfcLease" [ 1005.131955] env[62692]: } is ready. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1005.132207] env[62692]: DEBUG oslo_vmware.rw_handles [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1005.132207] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]526320df-b908-c34f-34ab-a4077bdec8a0" [ 1005.132207] env[62692]: _type = "HttpNfcLease" [ 1005.132207] env[62692]: }. {{(pid=62692) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1005.132935] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1f081a8-90d1-40e2-85ca-9acf5f1be1d1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.140569] env[62692]: DEBUG oslo_vmware.rw_handles [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c0dc5a-5295-2d56-afed-91e0e3dafbef/disk-0.vmdk from lease info. {{(pid=62692) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1005.140748] env[62692]: DEBUG oslo_vmware.rw_handles [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c0dc5a-5295-2d56-afed-91e0e3dafbef/disk-0.vmdk for reading. {{(pid=62692) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1005.231107] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ce13fbb8-9d54-4dbc-97d8-109d792f1486 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.249166] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd901703-1e93-42a1-a88a-4cc4ff790201 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "f6085887-a5d3-4096-be38-d843bb54304a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.948s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.453918] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.896s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.456837] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.667s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.457176] env[62692]: DEBUG nova.objects.instance [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lazy-loading 'resources' on Instance uuid f7664593-1b7e-495f-8d0e-9137d7f180cc {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.476027] env[62692]: INFO nova.scheduler.client.report [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Deleted allocations for instance 084ce29c-302a-45cc-b36f-0143eccee298 [ 1005.538033] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3eea435b-ef99-418b-aa1d-55133acea437 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "interface-668fb0ea-7eed-4198-943c-8f916ec11368-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.538033] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3eea435b-ef99-418b-aa1d-55133acea437 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "interface-668fb0ea-7eed-4198-943c-8f916ec11368-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.538033] env[62692]: DEBUG nova.objects.instance [None req-3eea435b-ef99-418b-aa1d-55133acea437 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lazy-loading 'flavor' on Instance uuid 668fb0ea-7eed-4198-943c-8f916ec11368 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.584784] env[62692]: DEBUG nova.compute.manager [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1005.910033] env[62692]: DEBUG nova.virt.hardware [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1005.910482] env[62692]: DEBUG nova.virt.hardware [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1005.910482] env[62692]: DEBUG nova.virt.hardware [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1005.910706] env[62692]: DEBUG nova.virt.hardware [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1005.910852] env[62692]: DEBUG nova.virt.hardware [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1005.911139] env[62692]: DEBUG nova.virt.hardware [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1005.911437] env[62692]: DEBUG nova.virt.hardware [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1005.911763] env[62692]: DEBUG nova.virt.hardware [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1005.911985] env[62692]: DEBUG nova.virt.hardware [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1005.912752] env[62692]: DEBUG nova.virt.hardware [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1005.913065] env[62692]: DEBUG nova.virt.hardware [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1005.914290] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7099752-c477-4fa8-a734-f331d3d6b88e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.924251] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9327943c-ffc4-40ca-8278-68b73194b3d1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.984131] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cd476325-4145-422d-830b-9607c70153f5 tempest-ServerRescueTestJSONUnderV235-968691958 tempest-ServerRescueTestJSONUnderV235-968691958-project-member] Lock "084ce29c-302a-45cc-b36f-0143eccee298" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.645s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.042331] env[62692]: DEBUG nova.objects.instance [None req-3eea435b-ef99-418b-aa1d-55133acea437 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lazy-loading 'pci_requests' on Instance uuid 668fb0ea-7eed-4198-943c-8f916ec11368 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1006.084607] env[62692]: DEBUG nova.network.neutron [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Successfully updated port: 9867c68f-a0ff-499d-9b91-47cbd25c4b69 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1006.286381] env[62692]: DEBUG nova.compute.manager [req-15f57c3a-bdfa-4d6d-911d-67cb1d8ee4b7 req-b7630d1c-b9cc-46f4-b19d-9dfbae40d433 service nova] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Received event network-vif-plugged-9867c68f-a0ff-499d-9b91-47cbd25c4b69 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1006.286703] env[62692]: DEBUG oslo_concurrency.lockutils [req-15f57c3a-bdfa-4d6d-911d-67cb1d8ee4b7 req-b7630d1c-b9cc-46f4-b19d-9dfbae40d433 service nova] Acquiring lock "f523c482-8365-47b8-9398-a56a98fcffe8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.287033] env[62692]: DEBUG oslo_concurrency.lockutils [req-15f57c3a-bdfa-4d6d-911d-67cb1d8ee4b7 req-b7630d1c-b9cc-46f4-b19d-9dfbae40d433 service nova] Lock "f523c482-8365-47b8-9398-a56a98fcffe8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.287282] env[62692]: DEBUG oslo_concurrency.lockutils [req-15f57c3a-bdfa-4d6d-911d-67cb1d8ee4b7 req-b7630d1c-b9cc-46f4-b19d-9dfbae40d433 service nova] Lock "f523c482-8365-47b8-9398-a56a98fcffe8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.287629] env[62692]: DEBUG nova.compute.manager [req-15f57c3a-bdfa-4d6d-911d-67cb1d8ee4b7 req-b7630d1c-b9cc-46f4-b19d-9dfbae40d433 service nova] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] No waiting events found dispatching network-vif-plugged-9867c68f-a0ff-499d-9b91-47cbd25c4b69 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1006.287820] env[62692]: WARNING nova.compute.manager [req-15f57c3a-bdfa-4d6d-911d-67cb1d8ee4b7 req-b7630d1c-b9cc-46f4-b19d-9dfbae40d433 service nova] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Received unexpected event network-vif-plugged-9867c68f-a0ff-499d-9b91-47cbd25c4b69 for instance with vm_state building and task_state spawning. [ 1006.336192] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7670d398-5473-4ca1-9ef3-30459f5e6e61 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.344568] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e193e372-14f6-4424-ab8d-86cede60b800 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.380923] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-716000d2-358a-4a3b-b3ea-e2e87b831850 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.390211] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-729ba0e8-b098-42ca-b7e9-a0af5c17470c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.407583] env[62692]: DEBUG nova.compute.provider_tree [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1006.548031] env[62692]: DEBUG nova.objects.base [None req-3eea435b-ef99-418b-aa1d-55133acea437 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Object Instance<668fb0ea-7eed-4198-943c-8f916ec11368> lazy-loaded attributes: flavor,pci_requests {{(pid=62692) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1006.548031] env[62692]: DEBUG nova.network.neutron [None req-3eea435b-ef99-418b-aa1d-55133acea437 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1006.589131] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Acquiring lock "refresh_cache-f523c482-8365-47b8-9398-a56a98fcffe8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.589131] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Acquired lock "refresh_cache-f523c482-8365-47b8-9398-a56a98fcffe8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.589131] env[62692]: DEBUG nova.network.neutron [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1006.642750] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3eea435b-ef99-418b-aa1d-55133acea437 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "interface-668fb0ea-7eed-4198-943c-8f916ec11368-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.105s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.911266] env[62692]: DEBUG nova.scheduler.client.report [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1006.939438] env[62692]: DEBUG nova.compute.manager [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1006.940507] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-492078f0-485c-4560-b7dc-4ec204a7d6d7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.130760] env[62692]: DEBUG nova.network.neutron [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1007.290800] env[62692]: DEBUG nova.network.neutron [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Updating instance_info_cache with network_info: [{"id": "9867c68f-a0ff-499d-9b91-47cbd25c4b69", "address": "fa:16:3e:9b:20:fd", "network": {"id": "a74c4d30-3568-4e50-98f5-d2fcc05ec728", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.220", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "989968e8796a4a70a4dbd045418a475f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9867c68f-a0", "ovs_interfaceid": "9867c68f-a0ff-499d-9b91-47cbd25c4b69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.421280] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.964s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.423557] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 23.370s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.454087] env[62692]: INFO nova.compute.manager [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] instance snapshotting [ 1007.457098] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba2b668f-c7df-41fc-8dda-f83ba953e124 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.461415] env[62692]: INFO nova.scheduler.client.report [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Deleted allocations for instance f7664593-1b7e-495f-8d0e-9137d7f180cc [ 1007.482401] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dfad114-befa-49fd-9dee-1536fdfc2e48 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.794714] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Releasing lock "refresh_cache-f523c482-8365-47b8-9398-a56a98fcffe8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.795142] env[62692]: DEBUG nova.compute.manager [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Instance network_info: |[{"id": "9867c68f-a0ff-499d-9b91-47cbd25c4b69", "address": "fa:16:3e:9b:20:fd", "network": {"id": "a74c4d30-3568-4e50-98f5-d2fcc05ec728", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.220", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "989968e8796a4a70a4dbd045418a475f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9867c68f-a0", "ovs_interfaceid": "9867c68f-a0ff-499d-9b91-47cbd25c4b69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1007.795651] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9b:20:fd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd8383707-f093-40a7-a5ba-31b0e07cac45', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9867c68f-a0ff-499d-9b91-47cbd25c4b69', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1007.804194] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Creating folder: Project (7aeb525d0d3644b5805b26a1a12fb450). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1007.804337] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f472f305-4e71-4d3e-9cad-588480fe1dc4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.815512] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Created folder: Project (7aeb525d0d3644b5805b26a1a12fb450) in parent group-v248868. [ 1007.815512] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Creating folder: Instances. Parent ref: group-v249018. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1007.815924] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7443e69b-cba1-49ba-a6c4-1f225a869616 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.826442] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Created folder: Instances in parent group-v249018. [ 1007.826707] env[62692]: DEBUG oslo.service.loopingcall [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1007.826942] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1007.827174] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dc91b8f6-cc55-4f2d-8594-221faa9f8996 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.850674] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1007.850674] env[62692]: value = "task-1141597" [ 1007.850674] env[62692]: _type = "Task" [ 1007.850674] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.860388] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141597, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.986105] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e321feec-967c-4b44-b3c6-656377d60734 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "f7664593-1b7e-495f-8d0e-9137d7f180cc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.424s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.999632] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Creating Snapshot of the VM instance {{(pid=62692) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1007.999988] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-ca2fba63-0ecf-4842-b4d1-35bf73a1930e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.007917] env[62692]: DEBUG oslo_vmware.api [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 1008.007917] env[62692]: value = "task-1141598" [ 1008.007917] env[62692]: _type = "Task" [ 1008.007917] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.017594] env[62692]: DEBUG oslo_vmware.api [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141598, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.188023] env[62692]: DEBUG oslo_concurrency.lockutils [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "b23d3484-4fc5-46b2-8e8c-bdaab087f046" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.188023] env[62692]: DEBUG oslo_concurrency.lockutils [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "b23d3484-4fc5-46b2-8e8c-bdaab087f046" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.188023] env[62692]: DEBUG oslo_concurrency.lockutils [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "b23d3484-4fc5-46b2-8e8c-bdaab087f046-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.188023] env[62692]: DEBUG oslo_concurrency.lockutils [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "b23d3484-4fc5-46b2-8e8c-bdaab087f046-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.188023] env[62692]: DEBUG oslo_concurrency.lockutils [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "b23d3484-4fc5-46b2-8e8c-bdaab087f046-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.190071] env[62692]: INFO nova.compute.manager [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Terminating instance [ 1008.193544] env[62692]: DEBUG nova.compute.manager [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1008.193738] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1008.194817] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e236ba0-3fcd-4fd6-8bef-917d523c25e8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.203671] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1008.203961] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-06412177-8fa2-4b91-a1fd-879b0f70327c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.210142] env[62692]: DEBUG oslo_vmware.api [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 1008.210142] env[62692]: value = "task-1141599" [ 1008.210142] env[62692]: _type = "Task" [ 1008.210142] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.219565] env[62692]: DEBUG oslo_vmware.api [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141599, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.362091] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141597, 'name': CreateVM_Task, 'duration_secs': 0.453162} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.362091] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1008.362306] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.362500] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.362881] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1008.363237] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d449be23-a709-4762-836a-4ce2c4758635 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.368248] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Waiting for the task: (returnval){ [ 1008.368248] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52cc2483-9482-7076-a6d0-ded2a3e729ed" [ 1008.368248] env[62692]: _type = "Task" [ 1008.368248] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.377680] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52cc2483-9482-7076-a6d0-ded2a3e729ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.443018] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Applying migration context for instance 3f6c9744-a6e3-43f3-8b6b-624ba178747d as it has an incoming, in-progress migration 4c7b8a13-dd0a-48d9-b1fe-523c53945e0c. Migration status is confirming {{(pid=62692) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1008.446121] env[62692]: INFO nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Updating resource usage from migration 4c7b8a13-dd0a-48d9-b1fe-523c53945e0c [ 1008.446500] env[62692]: INFO nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Updating resource usage from migration 91b03d62-ef95-4521-8e30-c12febaeacc3 [ 1008.475389] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 5a08fa8a-f9fe-4879-bb7b-baa04097df6c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.475556] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 9b8e830c-61b7-4dd2-8324-d3a96eec1465 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.475700] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance bb547773-d176-4c8e-a0fa-a374d5050b1b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.475831] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.475949] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance e2383a6a-3581-40fc-a0eb-6981acdbf54a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.476079] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 754b3c65-1e4b-49d2-8980-095d975edb01 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.476225] env[62692]: WARNING nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 26c1ff5b-9eda-4131-a4ea-d4511a2364b7 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1008.476342] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.476483] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance b23d3484-4fc5-46b2-8e8c-bdaab087f046 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.476625] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 668fb0ea-7eed-4198-943c-8f916ec11368 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.476749] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance f3e8be1c-d952-49b7-86a8-d0a6c56ee096 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.476862] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 16790505-cdf6-4937-8839-fa685a5f413e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.476974] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Migration 4c7b8a13-dd0a-48d9-b1fe-523c53945e0c is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1008.477098] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 3f6c9744-a6e3-43f3-8b6b-624ba178747d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.477210] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 250b9cc2-706e-4fbf-887c-1c80fc27a0ed actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.477326] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 7183ec4a-992d-48f1-8dda-7f499c2f4e1d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.477465] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance f6085887-a5d3-4096-be38-d843bb54304a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.477567] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance f523c482-8365-47b8-9398-a56a98fcffe8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.519025] env[62692]: DEBUG oslo_vmware.api [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141598, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.567548] env[62692]: DEBUG nova.compute.manager [req-3364b799-fa0c-4722-9a2d-3f599bdb05b6 req-a778d5c0-b742-4716-b699-5f69f7926bd6 service nova] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Received event network-changed-9867c68f-a0ff-499d-9b91-47cbd25c4b69 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1008.567673] env[62692]: DEBUG nova.compute.manager [req-3364b799-fa0c-4722-9a2d-3f599bdb05b6 req-a778d5c0-b742-4716-b699-5f69f7926bd6 service nova] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Refreshing instance network info cache due to event network-changed-9867c68f-a0ff-499d-9b91-47cbd25c4b69. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1008.567841] env[62692]: DEBUG oslo_concurrency.lockutils [req-3364b799-fa0c-4722-9a2d-3f599bdb05b6 req-a778d5c0-b742-4716-b699-5f69f7926bd6 service nova] Acquiring lock "refresh_cache-f523c482-8365-47b8-9398-a56a98fcffe8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.568024] env[62692]: DEBUG oslo_concurrency.lockutils [req-3364b799-fa0c-4722-9a2d-3f599bdb05b6 req-a778d5c0-b742-4716-b699-5f69f7926bd6 service nova] Acquired lock "refresh_cache-f523c482-8365-47b8-9398-a56a98fcffe8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.568187] env[62692]: DEBUG nova.network.neutron [req-3364b799-fa0c-4722-9a2d-3f599bdb05b6 req-a778d5c0-b742-4716-b699-5f69f7926bd6 service nova] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Refreshing network info cache for port 9867c68f-a0ff-499d-9b91-47cbd25c4b69 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1008.722441] env[62692]: DEBUG oslo_vmware.api [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141599, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.747231] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "interface-668fb0ea-7eed-4198-943c-8f916ec11368-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.747650] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "interface-668fb0ea-7eed-4198-943c-8f916ec11368-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.748097] env[62692]: DEBUG nova.objects.instance [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lazy-loading 'flavor' on Instance uuid 668fb0ea-7eed-4198-943c-8f916ec11368 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1008.878605] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52cc2483-9482-7076-a6d0-ded2a3e729ed, 'name': SearchDatastore_Task, 'duration_secs': 0.017188} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.878937] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.879185] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1008.879449] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.879621] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.879815] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1008.880138] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-16dd12c5-f97f-4241-808b-5a487a41aaa3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.888941] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1008.889190] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1008.889894] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2848c65-3047-4f47-b5fd-2b4f17452087 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.895402] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Waiting for the task: (returnval){ [ 1008.895402] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52b87b40-77ae-31bc-4833-05134e798a71" [ 1008.895402] env[62692]: _type = "Task" [ 1008.895402] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.903457] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52b87b40-77ae-31bc-4833-05134e798a71, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.904895] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9769e7f3-0a17-4a2a-ab0b-bccc49191252 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.905124] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9769e7f3-0a17-4a2a-ab0b-bccc49191252 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.980446] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 1b753882-9118-479c-aa3e-cbb8bdf2e086 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1008.980694] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Migration 91b03d62-ef95-4521-8e30-c12febaeacc3 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1008.980738] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance d99737d8-2eb0-40ee-b61e-6c736c84ea59 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1009.017685] env[62692]: DEBUG oslo_vmware.api [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141598, 'name': CreateSnapshot_Task, 'duration_secs': 0.840724} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.017974] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Created Snapshot of the VM instance {{(pid=62692) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1009.018729] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b108dd0f-0c73-40b2-b033-7196031ce8ff {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.228223] env[62692]: DEBUG oslo_vmware.api [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141599, 'name': PowerOffVM_Task, 'duration_secs': 0.639953} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.228766] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1009.229366] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1009.229580] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-729c3858-3e5d-4d73-9b44-abddfc76c8ea {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.302521] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1009.302809] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1009.302953] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Deleting the datastore file [datastore2] b23d3484-4fc5-46b2-8e8c-bdaab087f046 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1009.303243] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4b145c87-d8a6-4408-bf33-5ba3079c610c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.310576] env[62692]: DEBUG oslo_vmware.api [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 1009.310576] env[62692]: value = "task-1141601" [ 1009.310576] env[62692]: _type = "Task" [ 1009.310576] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.321159] env[62692]: DEBUG oslo_vmware.api [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141601, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.333208] env[62692]: DEBUG nova.objects.instance [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lazy-loading 'pci_requests' on Instance uuid 668fb0ea-7eed-4198-943c-8f916ec11368 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1009.388971] env[62692]: DEBUG nova.network.neutron [req-3364b799-fa0c-4722-9a2d-3f599bdb05b6 req-a778d5c0-b742-4716-b699-5f69f7926bd6 service nova] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Updated VIF entry in instance network info cache for port 9867c68f-a0ff-499d-9b91-47cbd25c4b69. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1009.389369] env[62692]: DEBUG nova.network.neutron [req-3364b799-fa0c-4722-9a2d-3f599bdb05b6 req-a778d5c0-b742-4716-b699-5f69f7926bd6 service nova] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Updating instance_info_cache with network_info: [{"id": "9867c68f-a0ff-499d-9b91-47cbd25c4b69", "address": "fa:16:3e:9b:20:fd", "network": {"id": "a74c4d30-3568-4e50-98f5-d2fcc05ec728", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.220", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "989968e8796a4a70a4dbd045418a475f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9867c68f-a0", "ovs_interfaceid": "9867c68f-a0ff-499d-9b91-47cbd25c4b69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.411474] env[62692]: INFO nova.compute.manager [None req-9769e7f3-0a17-4a2a-ab0b-bccc49191252 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Detaching volume 50c399be-767e-458d-bf5a-9c2395ab8e24 [ 1009.414188] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52b87b40-77ae-31bc-4833-05134e798a71, 'name': SearchDatastore_Task, 'duration_secs': 0.02758} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.415921] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44dd4269-8045-4531-bda9-c9d265d53120 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.426692] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Waiting for the task: (returnval){ [ 1009.426692] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52eb972f-fdb9-0ebc-1972-961a794a36a5" [ 1009.426692] env[62692]: _type = "Task" [ 1009.426692] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.437442] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52eb972f-fdb9-0ebc-1972-961a794a36a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.453868] env[62692]: INFO nova.virt.block_device [None req-9769e7f3-0a17-4a2a-ab0b-bccc49191252 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Attempting to driver detach volume 50c399be-767e-458d-bf5a-9c2395ab8e24 from mountpoint /dev/sdb [ 1009.453868] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-9769e7f3-0a17-4a2a-ab0b-bccc49191252 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Volume detach. Driver type: vmdk {{(pid=62692) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1009.453868] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-9769e7f3-0a17-4a2a-ab0b-bccc49191252 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-248997', 'volume_id': '50c399be-767e-458d-bf5a-9c2395ab8e24', 'name': 'volume-50c399be-767e-458d-bf5a-9c2395ab8e24', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19', 'attached_at': '', 'detached_at': '', 'volume_id': '50c399be-767e-458d-bf5a-9c2395ab8e24', 'serial': '50c399be-767e-458d-bf5a-9c2395ab8e24'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1009.454365] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db17b245-3e62-4dc9-a60a-3972f16ea95c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.476697] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da14d0bc-6901-4d48-a308-dbd74f6d4a07 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.483631] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance e8f645da-3b4a-4f21-9a17-aebdc8d112df has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1009.483925] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Total usable vcpus: 48, total allocated vcpus: 18 {{(pid=62692) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1009.484099] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4032MB phys_disk=200GB used_disk=18GB total_vcpus=48 used_vcpus=18 pci_stats=[] {{(pid=62692) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1009.487979] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b724142-e9de-463a-8b5e-01dbb858e72f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.513000] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bfb97f4-b234-4fca-9779-e2679a5b0a76 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.528654] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-9769e7f3-0a17-4a2a-ab0b-bccc49191252 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] The volume has not been displaced from its original location: [datastore2] volume-50c399be-767e-458d-bf5a-9c2395ab8e24/volume-50c399be-767e-458d-bf5a-9c2395ab8e24.vmdk. No consolidation needed. {{(pid=62692) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1009.534303] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-9769e7f3-0a17-4a2a-ab0b-bccc49191252 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Reconfiguring VM instance instance-00000040 to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1009.544037] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Creating linked-clone VM from snapshot {{(pid=62692) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1009.544204] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac954f4b-e290-4633-8f25-15f25373a3a9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.558738] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a7d9b427-1ad7-4640-b3bd-21dce38f090e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.567404] env[62692]: DEBUG oslo_vmware.api [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 1009.567404] env[62692]: value = "task-1141602" [ 1009.567404] env[62692]: _type = "Task" [ 1009.567404] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.568799] env[62692]: DEBUG oslo_vmware.api [None req-9769e7f3-0a17-4a2a-ab0b-bccc49191252 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 1009.568799] env[62692]: value = "task-1141603" [ 1009.568799] env[62692]: _type = "Task" [ 1009.568799] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.584206] env[62692]: DEBUG oslo_vmware.api [None req-9769e7f3-0a17-4a2a-ab0b-bccc49191252 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141603, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.592033] env[62692]: DEBUG oslo_vmware.api [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141602, 'name': CloneVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.820382] env[62692]: DEBUG oslo_vmware.api [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141601, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.384025} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.824118] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1009.824350] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1009.824534] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1009.824710] env[62692]: INFO nova.compute.manager [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Took 1.63 seconds to destroy the instance on the hypervisor. [ 1009.824958] env[62692]: DEBUG oslo.service.loopingcall [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1009.826278] env[62692]: DEBUG nova.compute.manager [-] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1009.826278] env[62692]: DEBUG nova.network.neutron [-] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1009.837862] env[62692]: DEBUG nova.objects.base [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Object Instance<668fb0ea-7eed-4198-943c-8f916ec11368> lazy-loaded attributes: flavor,pci_requests {{(pid=62692) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1009.838101] env[62692]: DEBUG nova.network.neutron [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1009.852930] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda5000c-c2b5-45ac-a08d-971dcd4dd74c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.860031] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f12bad82-6489-4cb9-8348-6ae7e18cdb56 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.891919] env[62692]: DEBUG nova.policy [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a251ef5f6cb143178928babc75a0fef0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ba97a775e99a463c845518b0df8cd3e2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1009.895645] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17a1566c-7e24-443b-b184-35cde73aa59b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.898303] env[62692]: DEBUG oslo_concurrency.lockutils [req-3364b799-fa0c-4722-9a2d-3f599bdb05b6 req-a778d5c0-b742-4716-b699-5f69f7926bd6 service nova] Releasing lock "refresh_cache-f523c482-8365-47b8-9398-a56a98fcffe8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.904373] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4230cdb-6189-4e40-9160-6396bceb405d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.919470] env[62692]: DEBUG nova.compute.provider_tree [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1009.938536] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52eb972f-fdb9-0ebc-1972-961a794a36a5, 'name': SearchDatastore_Task, 'duration_secs': 0.018233} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.940689] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.940689] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] f523c482-8365-47b8-9398-a56a98fcffe8/f523c482-8365-47b8-9398-a56a98fcffe8.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1009.940689] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-84274063-56d2-4354-955f-1af62df285a2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.949851] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Waiting for the task: (returnval){ [ 1009.949851] env[62692]: value = "task-1141604" [ 1009.949851] env[62692]: _type = "Task" [ 1009.949851] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.958697] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Task: {'id': task-1141604, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.082224] env[62692]: DEBUG oslo_vmware.api [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141602, 'name': CloneVM_Task} progress is 94%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.088923] env[62692]: DEBUG oslo_vmware.api [None req-9769e7f3-0a17-4a2a-ab0b-bccc49191252 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141603, 'name': ReconfigVM_Task, 'duration_secs': 0.253875} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.089320] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-9769e7f3-0a17-4a2a-ab0b-bccc49191252 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Reconfigured VM instance instance-00000040 to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1010.094064] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b20885a-fe27-4475-bcdb-2443595b877a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.110537] env[62692]: DEBUG oslo_vmware.api [None req-9769e7f3-0a17-4a2a-ab0b-bccc49191252 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 1010.110537] env[62692]: value = "task-1141605" [ 1010.110537] env[62692]: _type = "Task" [ 1010.110537] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.121276] env[62692]: DEBUG oslo_vmware.api [None req-9769e7f3-0a17-4a2a-ab0b-bccc49191252 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141605, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.218733] env[62692]: DEBUG nova.network.neutron [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Successfully created port: e1a45cde-8d94-464d-8c3e-e3f94b62f9e6 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1010.423446] env[62692]: DEBUG nova.scheduler.client.report [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1010.462851] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Task: {'id': task-1141604, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.505555] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "f3e8be1c-d952-49b7-86a8-d0a6c56ee096" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.505872] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "f3e8be1c-d952-49b7-86a8-d0a6c56ee096" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.506102] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "f3e8be1c-d952-49b7-86a8-d0a6c56ee096-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.506294] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "f3e8be1c-d952-49b7-86a8-d0a6c56ee096-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.506485] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "f3e8be1c-d952-49b7-86a8-d0a6c56ee096-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.509936] env[62692]: INFO nova.compute.manager [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Terminating instance [ 1010.514931] env[62692]: DEBUG nova.compute.manager [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1010.515199] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1010.516106] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0084c1e1-cfcc-440c-ad17-9718cef4adf2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.525242] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1010.525612] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fc8ba20a-7679-4ba5-ae75-9e9a11d032d8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.535136] env[62692]: DEBUG oslo_vmware.api [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1010.535136] env[62692]: value = "task-1141606" [ 1010.535136] env[62692]: _type = "Task" [ 1010.535136] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.547296] env[62692]: DEBUG oslo_vmware.api [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141606, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.580337] env[62692]: DEBUG oslo_vmware.api [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141602, 'name': CloneVM_Task} progress is 94%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.621492] env[62692]: DEBUG oslo_vmware.api [None req-9769e7f3-0a17-4a2a-ab0b-bccc49191252 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141605, 'name': ReconfigVM_Task, 'duration_secs': 0.168491} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.621836] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-9769e7f3-0a17-4a2a-ab0b-bccc49191252 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-248997', 'volume_id': '50c399be-767e-458d-bf5a-9c2395ab8e24', 'name': 'volume-50c399be-767e-458d-bf5a-9c2395ab8e24', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19', 'attached_at': '', 'detached_at': '', 'volume_id': '50c399be-767e-458d-bf5a-9c2395ab8e24', 'serial': '50c399be-767e-458d-bf5a-9c2395ab8e24'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1010.673073] env[62692]: DEBUG nova.network.neutron [-] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.832574] env[62692]: DEBUG nova.compute.manager [req-beb36e8d-85f1-4fb2-99b1-4e518309b53c req-28f9901a-db52-48af-a415-0b0084333f57 service nova] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Received event network-vif-deleted-7cf21363-a84c-4404-b293-ccee3ea80385 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1010.928386] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62692) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1010.928858] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.505s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.929079] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.459s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.930770] env[62692]: INFO nova.compute.claims [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1010.960818] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Task: {'id': task-1141604, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.761578} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.961142] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] f523c482-8365-47b8-9398-a56a98fcffe8/f523c482-8365-47b8-9398-a56a98fcffe8.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1010.961466] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1010.962077] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6f65b83d-40d2-438e-82e8-a42e8a542945 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.970612] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Waiting for the task: (returnval){ [ 1010.970612] env[62692]: value = "task-1141607" [ 1010.970612] env[62692]: _type = "Task" [ 1010.970612] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.979958] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Task: {'id': task-1141607, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.046120] env[62692]: DEBUG oslo_vmware.api [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141606, 'name': PowerOffVM_Task, 'duration_secs': 0.224715} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.046403] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1011.046994] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1011.046994] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-788844e2-e51c-4867-af05-467c73cc4827 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.080359] env[62692]: DEBUG oslo_vmware.api [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141602, 'name': CloneVM_Task, 'duration_secs': 1.457595} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.080774] env[62692]: INFO nova.virt.vmwareapi.vmops [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Created linked-clone VM from snapshot [ 1011.081539] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95828e0e-2ed2-44a2-bfa2-f9350ad70f6d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.090754] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Uploading image bc25caa7-5b53-4670-b2b7-9a61d6e56819 {{(pid=62692) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1011.105764] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Destroying the VM {{(pid=62692) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1011.106179] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c027456b-1c7f-4c41-87d1-a38dcc332ae1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.113011] env[62692]: DEBUG oslo_vmware.api [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 1011.113011] env[62692]: value = "task-1141609" [ 1011.113011] env[62692]: _type = "Task" [ 1011.113011] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.122421] env[62692]: DEBUG oslo_vmware.api [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141609, 'name': Destroy_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.134051] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1011.134332] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1011.134534] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Deleting the datastore file [datastore1] f3e8be1c-d952-49b7-86a8-d0a6c56ee096 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1011.134859] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6f2d3c2c-d983-418b-bf10-215e1b073b3a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.142977] env[62692]: DEBUG oslo_vmware.api [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1011.142977] env[62692]: value = "task-1141610" [ 1011.142977] env[62692]: _type = "Task" [ 1011.142977] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.152429] env[62692]: DEBUG oslo_vmware.api [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141610, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.174095] env[62692]: DEBUG nova.objects.instance [None req-9769e7f3-0a17-4a2a-ab0b-bccc49191252 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lazy-loading 'flavor' on Instance uuid 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1011.175847] env[62692]: INFO nova.compute.manager [-] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Took 1.35 seconds to deallocate network for instance. [ 1011.479663] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Task: {'id': task-1141607, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07351} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.479964] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1011.480808] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33c85e2d-43fe-492b-b1bc-64adc14a97db {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.503416] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] f523c482-8365-47b8-9398-a56a98fcffe8/f523c482-8365-47b8-9398-a56a98fcffe8.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1011.503715] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-40b54563-855f-4a98-9a5c-5b57d7183ca4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.523912] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Waiting for the task: (returnval){ [ 1011.523912] env[62692]: value = "task-1141611" [ 1011.523912] env[62692]: _type = "Task" [ 1011.523912] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.532498] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Task: {'id': task-1141611, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.624148] env[62692]: DEBUG oslo_vmware.api [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141609, 'name': Destroy_Task} progress is 33%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.653380] env[62692]: DEBUG oslo_vmware.api [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141610, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.254812} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.653817] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1011.654082] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1011.654320] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1011.654546] env[62692]: INFO nova.compute.manager [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1011.654842] env[62692]: DEBUG oslo.service.loopingcall [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1011.655330] env[62692]: DEBUG nova.compute.manager [-] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1011.655480] env[62692]: DEBUG nova.network.neutron [-] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1011.681921] env[62692]: DEBUG oslo_concurrency.lockutils [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.768900] env[62692]: DEBUG nova.compute.manager [req-24596f7c-dec3-4878-b413-e8661e760127 req-68426b03-ef4d-4b90-89c6-ba7a45a9bffb service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Received event network-vif-plugged-e1a45cde-8d94-464d-8c3e-e3f94b62f9e6 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1011.769115] env[62692]: DEBUG oslo_concurrency.lockutils [req-24596f7c-dec3-4878-b413-e8661e760127 req-68426b03-ef4d-4b90-89c6-ba7a45a9bffb service nova] Acquiring lock "668fb0ea-7eed-4198-943c-8f916ec11368-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.769340] env[62692]: DEBUG oslo_concurrency.lockutils [req-24596f7c-dec3-4878-b413-e8661e760127 req-68426b03-ef4d-4b90-89c6-ba7a45a9bffb service nova] Lock "668fb0ea-7eed-4198-943c-8f916ec11368-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.769564] env[62692]: DEBUG oslo_concurrency.lockutils [req-24596f7c-dec3-4878-b413-e8661e760127 req-68426b03-ef4d-4b90-89c6-ba7a45a9bffb service nova] Lock "668fb0ea-7eed-4198-943c-8f916ec11368-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.769737] env[62692]: DEBUG nova.compute.manager [req-24596f7c-dec3-4878-b413-e8661e760127 req-68426b03-ef4d-4b90-89c6-ba7a45a9bffb service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] No waiting events found dispatching network-vif-plugged-e1a45cde-8d94-464d-8c3e-e3f94b62f9e6 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1011.769920] env[62692]: WARNING nova.compute.manager [req-24596f7c-dec3-4878-b413-e8661e760127 req-68426b03-ef4d-4b90-89c6-ba7a45a9bffb service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Received unexpected event network-vif-plugged-e1a45cde-8d94-464d-8c3e-e3f94b62f9e6 for instance with vm_state active and task_state None. [ 1011.800252] env[62692]: DEBUG nova.network.neutron [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Successfully updated port: e1a45cde-8d94-464d-8c3e-e3f94b62f9e6 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1012.044387] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Task: {'id': task-1141611, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.127651] env[62692]: DEBUG oslo_vmware.api [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141609, 'name': Destroy_Task, 'duration_secs': 0.54573} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.127924] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Destroyed the VM [ 1012.131824] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Deleting Snapshot of the VM instance {{(pid=62692) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1012.134685] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-6f2d79d1-e54c-4bed-995c-4ad021ab838b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.142113] env[62692]: DEBUG oslo_vmware.api [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 1012.142113] env[62692]: value = "task-1141612" [ 1012.142113] env[62692]: _type = "Task" [ 1012.142113] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.150815] env[62692]: DEBUG oslo_vmware.api [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141612, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.184773] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9769e7f3-0a17-4a2a-ab0b-bccc49191252 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.279s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.272050] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1d4e49d-f89b-4708-a8b9-04be4830df17 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.281343] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb9c9a92-4f54-4106-93df-d712806242d0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.315470] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "refresh_cache-668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.315624] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "refresh_cache-668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.315841] env[62692]: DEBUG nova.network.neutron [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1012.317882] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e95eb3f-e2f8-42ad-9c64-2f3c4444dc5b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.327433] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7728156-9577-4ae8-a413-2b51f0009c79 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.345602] env[62692]: DEBUG nova.compute.provider_tree [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1012.539190] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Task: {'id': task-1141611, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.652141] env[62692]: DEBUG oslo_vmware.api [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141612, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.684863] env[62692]: DEBUG nova.network.neutron [-] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.847979] env[62692]: DEBUG nova.scheduler.client.report [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1012.873573] env[62692]: WARNING nova.network.neutron [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] b34af32c-59b7-4de4-85ef-08bd67c1e2a2 already exists in list: networks containing: ['b34af32c-59b7-4de4-85ef-08bd67c1e2a2']. ignoring it [ 1013.040251] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Task: {'id': task-1141611, 'name': ReconfigVM_Task, 'duration_secs': 1.433473} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.040635] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Reconfigured VM instance instance-00000057 to attach disk [datastore2] f523c482-8365-47b8-9398-a56a98fcffe8/f523c482-8365-47b8-9398-a56a98fcffe8.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1013.041145] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-42710cc9-1094-4611-9f8e-31382e4b4a81 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.048841] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Waiting for the task: (returnval){ [ 1013.048841] env[62692]: value = "task-1141613" [ 1013.048841] env[62692]: _type = "Task" [ 1013.048841] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.054423] env[62692]: DEBUG nova.compute.manager [req-9680f0a7-490d-4ed5-86ab-4ef1fbbff9d3 req-1ed66fed-1a83-46c2-bd65-b4ab1616137a service nova] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Received event network-vif-deleted-1b3f2a08-aff4-4b7e-8b44-bc9033d93162 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1013.063226] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Task: {'id': task-1141613, 'name': Rename_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.152185] env[62692]: DEBUG oslo_vmware.api [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141612, 'name': RemoveSnapshot_Task, 'duration_secs': 0.803343} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.154991] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Deleted Snapshot of the VM instance {{(pid=62692) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1013.187894] env[62692]: INFO nova.compute.manager [-] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Took 1.53 seconds to deallocate network for instance. [ 1013.251870] env[62692]: DEBUG nova.network.neutron [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Updating instance_info_cache with network_info: [{"id": "2deb4259-2f64-4f7c-9289-eba76375335a", "address": "fa:16:3e:bd:3f:b3", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.135", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2deb4259-2f", "ovs_interfaceid": "2deb4259-2f64-4f7c-9289-eba76375335a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e1a45cde-8d94-464d-8c3e-e3f94b62f9e6", "address": "fa:16:3e:aa:dc:ce", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1a45cde-8d", "ovs_interfaceid": "e1a45cde-8d94-464d-8c3e-e3f94b62f9e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.354209] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.424s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.354209] env[62692]: DEBUG nova.compute.manager [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1013.357403] env[62692]: DEBUG oslo_concurrency.lockutils [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.508s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.357639] env[62692]: DEBUG oslo_concurrency.lockutils [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.359937] env[62692]: DEBUG oslo_concurrency.lockutils [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 12.858s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.382748] env[62692]: INFO nova.scheduler.client.report [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Deleted allocations for instance 26c1ff5b-9eda-4131-a4ea-d4511a2364b7 [ 1013.559870] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Task: {'id': task-1141613, 'name': Rename_Task, 'duration_secs': 0.164753} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.560163] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1013.560457] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-50e7c2d0-6476-4fc3-b02a-8091454b7918 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.567017] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Waiting for the task: (returnval){ [ 1013.567017] env[62692]: value = "task-1141614" [ 1013.567017] env[62692]: _type = "Task" [ 1013.567017] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.575136] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Task: {'id': task-1141614, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.598343] env[62692]: DEBUG oslo_vmware.rw_handles [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c0dc5a-5295-2d56-afed-91e0e3dafbef/disk-0.vmdk. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1013.599389] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a15d0fd-d2ea-4ef5-9e9f-601061b32270 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.606038] env[62692]: DEBUG oslo_vmware.rw_handles [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c0dc5a-5295-2d56-afed-91e0e3dafbef/disk-0.vmdk is in state: ready. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1013.606219] env[62692]: ERROR oslo_vmware.rw_handles [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c0dc5a-5295-2d56-afed-91e0e3dafbef/disk-0.vmdk due to incomplete transfer. [ 1013.606452] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-571538cb-33c9-4a75-b87e-f31109750327 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.613930] env[62692]: DEBUG oslo_vmware.rw_handles [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c0dc5a-5295-2d56-afed-91e0e3dafbef/disk-0.vmdk. {{(pid=62692) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1013.614157] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Uploaded image 01209948-5245-453f-bb56-fb80a8ea2d67 to the Glance image server {{(pid=62692) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1013.616572] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Destroying the VM {{(pid=62692) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1013.616827] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a1e6feda-e839-40a5-9861-3318c4a1e266 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.623861] env[62692]: DEBUG oslo_vmware.api [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1013.623861] env[62692]: value = "task-1141615" [ 1013.623861] env[62692]: _type = "Task" [ 1013.623861] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.634429] env[62692]: DEBUG oslo_vmware.api [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141615, 'name': Destroy_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.660899] env[62692]: WARNING nova.compute.manager [None req-a9373ede-221c-43a5-bba4-588af21aef42 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Image not found during snapshot: nova.exception.ImageNotFound: Image bc25caa7-5b53-4670-b2b7-9a61d6e56819 could not be found. [ 1013.680591] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.680849] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.681083] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.681344] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.681545] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.683605] env[62692]: INFO nova.compute.manager [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Terminating instance [ 1013.686062] env[62692]: DEBUG nova.compute.manager [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1013.686264] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1013.687106] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7685b187-b263-41b5-983f-6a985c3c15c7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.695245] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1013.695483] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-58a41489-d1ec-425b-9711-a4aa7a52779b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.698097] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.703928] env[62692]: DEBUG oslo_vmware.api [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 1013.703928] env[62692]: value = "task-1141616" [ 1013.703928] env[62692]: _type = "Task" [ 1013.703928] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.712293] env[62692]: DEBUG oslo_vmware.api [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141616, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.757132] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "refresh_cache-668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.757967] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.758190] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.759044] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e60957ca-b932-4c60-9c59-9e68874fc231 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.778282] env[62692]: DEBUG nova.virt.hardware [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1013.778636] env[62692]: DEBUG nova.virt.hardware [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1013.778911] env[62692]: DEBUG nova.virt.hardware [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1013.779243] env[62692]: DEBUG nova.virt.hardware [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1013.779444] env[62692]: DEBUG nova.virt.hardware [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1013.779616] env[62692]: DEBUG nova.virt.hardware [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1013.779838] env[62692]: DEBUG nova.virt.hardware [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1013.780014] env[62692]: DEBUG nova.virt.hardware [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1013.780207] env[62692]: DEBUG nova.virt.hardware [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1013.780380] env[62692]: DEBUG nova.virt.hardware [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1013.780561] env[62692]: DEBUG nova.virt.hardware [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1013.786817] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Reconfiguring VM to attach interface {{(pid=62692) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1013.787464] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b9dcf8e7-a61f-4c2e-bc94-1c5636845f6b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.806768] env[62692]: DEBUG oslo_vmware.api [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1013.806768] env[62692]: value = "task-1141617" [ 1013.806768] env[62692]: _type = "Task" [ 1013.806768] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.815887] env[62692]: DEBUG oslo_vmware.api [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141617, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.863067] env[62692]: DEBUG nova.compute.utils [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1013.866532] env[62692]: INFO nova.compute.claims [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1013.870419] env[62692]: DEBUG nova.compute.manager [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1013.870529] env[62692]: DEBUG nova.network.neutron [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1013.890416] env[62692]: DEBUG oslo_concurrency.lockutils [None req-736323fa-244c-47ff-b821-8b4c609cd13d tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "26c1ff5b-9eda-4131-a4ea-d4511a2364b7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.985s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.931376] env[62692]: DEBUG nova.policy [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5f4f8dca6b6549a7b3aac06e3e64ba08', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f26b93a4a43b45e8a11f3793996c8496', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1014.021165] env[62692]: DEBUG nova.compute.manager [req-10dbfb19-e4c1-40b5-bb97-7a591a4a5c66 req-df328f95-360c-4c52-9dbb-f0077fe717a2 service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Received event network-changed-e1a45cde-8d94-464d-8c3e-e3f94b62f9e6 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1014.021165] env[62692]: DEBUG nova.compute.manager [req-10dbfb19-e4c1-40b5-bb97-7a591a4a5c66 req-df328f95-360c-4c52-9dbb-f0077fe717a2 service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Refreshing instance network info cache due to event network-changed-e1a45cde-8d94-464d-8c3e-e3f94b62f9e6. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1014.021165] env[62692]: DEBUG oslo_concurrency.lockutils [req-10dbfb19-e4c1-40b5-bb97-7a591a4a5c66 req-df328f95-360c-4c52-9dbb-f0077fe717a2 service nova] Acquiring lock "refresh_cache-668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.021165] env[62692]: DEBUG oslo_concurrency.lockutils [req-10dbfb19-e4c1-40b5-bb97-7a591a4a5c66 req-df328f95-360c-4c52-9dbb-f0077fe717a2 service nova] Acquired lock "refresh_cache-668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.022123] env[62692]: DEBUG nova.network.neutron [req-10dbfb19-e4c1-40b5-bb97-7a591a4a5c66 req-df328f95-360c-4c52-9dbb-f0077fe717a2 service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Refreshing network info cache for port e1a45cde-8d94-464d-8c3e-e3f94b62f9e6 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1014.080115] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Task: {'id': task-1141614, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.136842] env[62692]: DEBUG oslo_vmware.api [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141615, 'name': Destroy_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.218478] env[62692]: DEBUG oslo_vmware.api [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141616, 'name': PowerOffVM_Task, 'duration_secs': 0.472855} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.218478] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1014.218478] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1014.218478] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-56063466-14d7-4e63-a0b6-a58bebfc8484 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.255542] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "f6085887-a5d3-4096-be38-d843bb54304a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.255542] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "f6085887-a5d3-4096-be38-d843bb54304a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.255542] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "f6085887-a5d3-4096-be38-d843bb54304a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.255542] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "f6085887-a5d3-4096-be38-d843bb54304a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.255542] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "f6085887-a5d3-4096-be38-d843bb54304a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.256943] env[62692]: INFO nova.compute.manager [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Terminating instance [ 1014.259111] env[62692]: DEBUG nova.compute.manager [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1014.259642] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1014.260835] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-658f8382-7ff8-40de-ac15-2570af59969d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.269509] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1014.269943] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d7d84cb4-1406-46db-b6af-f4e7b6dc3ffd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.279750] env[62692]: DEBUG oslo_vmware.api [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 1014.279750] env[62692]: value = "task-1141619" [ 1014.279750] env[62692]: _type = "Task" [ 1014.279750] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.283160] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1014.284020] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1014.284338] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Deleting the datastore file [datastore2] 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1014.288503] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7af0c94e-d612-4131-ba99-b9cc4302d746 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.291648] env[62692]: DEBUG oslo_vmware.api [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141619, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.296893] env[62692]: DEBUG oslo_vmware.api [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 1014.296893] env[62692]: value = "task-1141620" [ 1014.296893] env[62692]: _type = "Task" [ 1014.296893] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.306640] env[62692]: DEBUG oslo_vmware.api [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141620, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.316177] env[62692]: DEBUG oslo_vmware.api [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141617, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.317031] env[62692]: DEBUG nova.network.neutron [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Successfully created port: c43f9978-4cfd-467d-86e0-4c8acfefc821 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1014.375775] env[62692]: DEBUG nova.compute.manager [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1014.377070] env[62692]: INFO nova.compute.resource_tracker [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Updating resource usage from migration 91b03d62-ef95-4521-8e30-c12febaeacc3 [ 1014.579581] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Task: {'id': task-1141614, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.637652] env[62692]: DEBUG oslo_vmware.api [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141615, 'name': Destroy_Task, 'duration_secs': 0.944348} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.640463] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Destroyed the VM [ 1014.640650] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Deleting Snapshot of the VM instance {{(pid=62692) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1014.640933] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5cf25e2f-0148-44ec-8f81-437ada92bf42 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.652214] env[62692]: DEBUG oslo_vmware.api [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1014.652214] env[62692]: value = "task-1141621" [ 1014.652214] env[62692]: _type = "Task" [ 1014.652214] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.663766] env[62692]: DEBUG oslo_vmware.api [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141621, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.702698] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-175a0c14-534e-4065-a23c-4186119ec349 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.710813] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07fa52c7-d696-4684-8dfa-77fafdb5e679 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.749963] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7633b7bd-8d4d-48c0-a965-2ecbd7b17a98 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.759377] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1f67972-1cd9-4f99-a774-56bd7c181b12 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.775610] env[62692]: DEBUG nova.compute.provider_tree [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1014.790310] env[62692]: DEBUG oslo_vmware.api [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141619, 'name': PowerOffVM_Task, 'duration_secs': 0.204634} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.790779] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1014.790866] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1014.791253] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-260afad0-47b3-4234-bf83-4beb62b34373 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.796818] env[62692]: DEBUG nova.network.neutron [req-10dbfb19-e4c1-40b5-bb97-7a591a4a5c66 req-df328f95-360c-4c52-9dbb-f0077fe717a2 service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Updated VIF entry in instance network info cache for port e1a45cde-8d94-464d-8c3e-e3f94b62f9e6. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1014.797251] env[62692]: DEBUG nova.network.neutron [req-10dbfb19-e4c1-40b5-bb97-7a591a4a5c66 req-df328f95-360c-4c52-9dbb-f0077fe717a2 service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Updating instance_info_cache with network_info: [{"id": "2deb4259-2f64-4f7c-9289-eba76375335a", "address": "fa:16:3e:bd:3f:b3", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.135", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2deb4259-2f", "ovs_interfaceid": "2deb4259-2f64-4f7c-9289-eba76375335a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e1a45cde-8d94-464d-8c3e-e3f94b62f9e6", "address": "fa:16:3e:aa:dc:ce", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1a45cde-8d", "ovs_interfaceid": "e1a45cde-8d94-464d-8c3e-e3f94b62f9e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.808716] env[62692]: DEBUG oslo_vmware.api [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141620, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187465} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.812599] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1014.812800] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1014.812987] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1014.813213] env[62692]: INFO nova.compute.manager [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1014.813467] env[62692]: DEBUG oslo.service.loopingcall [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1014.814270] env[62692]: DEBUG nova.compute.manager [-] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1014.814363] env[62692]: DEBUG nova.network.neutron [-] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1014.821810] env[62692]: DEBUG oslo_vmware.api [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141617, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.859701] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1014.860219] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1014.860445] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Deleting the datastore file [datastore1] f6085887-a5d3-4096-be38-d843bb54304a {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1014.860721] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e15c9b72-f6e3-4fd9-b67d-73019b8b613f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.872552] env[62692]: DEBUG oslo_vmware.api [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for the task: (returnval){ [ 1014.872552] env[62692]: value = "task-1141623" [ 1014.872552] env[62692]: _type = "Task" [ 1014.872552] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.889021] env[62692]: DEBUG oslo_vmware.api [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141623, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.082068] env[62692]: DEBUG oslo_vmware.api [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Task: {'id': task-1141614, 'name': PowerOnVM_Task, 'duration_secs': 1.300735} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.084839] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1015.084839] env[62692]: INFO nova.compute.manager [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Took 9.50 seconds to spawn the instance on the hypervisor. [ 1015.084839] env[62692]: DEBUG nova.compute.manager [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1015.084839] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-023acc89-2292-4202-86d5-1632e300b779 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.162595] env[62692]: DEBUG oslo_vmware.api [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141621, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.278978] env[62692]: DEBUG nova.scheduler.client.report [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1015.300845] env[62692]: DEBUG oslo_concurrency.lockutils [req-10dbfb19-e4c1-40b5-bb97-7a591a4a5c66 req-df328f95-360c-4c52-9dbb-f0077fe717a2 service nova] Releasing lock "refresh_cache-668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.318827] env[62692]: DEBUG oslo_vmware.api [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141617, 'name': ReconfigVM_Task, 'duration_secs': 1.071521} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.319382] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.319643] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Reconfigured VM to attach interface {{(pid=62692) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1015.385794] env[62692]: DEBUG nova.compute.manager [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1015.387907] env[62692]: DEBUG oslo_vmware.api [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Task: {'id': task-1141623, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.282374} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.388446] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "9b8e830c-61b7-4dd2-8324-d3a96eec1465" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.388695] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "9b8e830c-61b7-4dd2-8324-d3a96eec1465" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.388909] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "9b8e830c-61b7-4dd2-8324-d3a96eec1465-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.389205] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "9b8e830c-61b7-4dd2-8324-d3a96eec1465-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.389400] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "9b8e830c-61b7-4dd2-8324-d3a96eec1465-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.391373] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1015.391589] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1015.391875] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1015.395994] env[62692]: INFO nova.compute.manager [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1015.396259] env[62692]: DEBUG oslo.service.loopingcall [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1015.396740] env[62692]: INFO nova.compute.manager [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Terminating instance [ 1015.398233] env[62692]: DEBUG nova.compute.manager [-] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1015.398335] env[62692]: DEBUG nova.network.neutron [-] [instance: f6085887-a5d3-4096-be38-d843bb54304a] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1015.400265] env[62692]: DEBUG nova.compute.manager [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1015.400684] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1015.401290] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcb52aad-7214-4147-86b2-ab9cf325a149 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.409201] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1015.411071] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-73148136-a276-4600-bbc6-3eeb4ade875f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.418160] env[62692]: DEBUG oslo_vmware.api [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 1015.418160] env[62692]: value = "task-1141624" [ 1015.418160] env[62692]: _type = "Task" [ 1015.418160] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.421233] env[62692]: DEBUG nova.virt.hardware [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1015.421480] env[62692]: DEBUG nova.virt.hardware [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1015.421749] env[62692]: DEBUG nova.virt.hardware [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1015.421913] env[62692]: DEBUG nova.virt.hardware [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1015.422113] env[62692]: DEBUG nova.virt.hardware [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1015.422303] env[62692]: DEBUG nova.virt.hardware [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1015.422578] env[62692]: DEBUG nova.virt.hardware [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1015.422754] env[62692]: DEBUG nova.virt.hardware [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1015.422932] env[62692]: DEBUG nova.virt.hardware [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1015.423111] env[62692]: DEBUG nova.virt.hardware [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1015.423578] env[62692]: DEBUG nova.virt.hardware [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1015.427099] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37e5ea43-67f5-4926-a99a-03ab8a3d9a92 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.434965] env[62692]: DEBUG oslo_vmware.api [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141624, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.438251] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-438ddf0f-a5a9-476b-be45-d4c451b40712 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.604145] env[62692]: INFO nova.compute.manager [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Took 37.35 seconds to build instance. [ 1015.663761] env[62692]: DEBUG oslo_vmware.api [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141621, 'name': RemoveSnapshot_Task, 'duration_secs': 0.745659} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.664709] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Deleted Snapshot of the VM instance {{(pid=62692) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1015.665035] env[62692]: DEBUG nova.compute.manager [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1015.665836] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5107681-bda1-45f6-b8b8-9ba39c1d2110 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.784981] env[62692]: DEBUG oslo_concurrency.lockutils [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.425s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.785269] env[62692]: INFO nova.compute.manager [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Migrating [ 1015.792194] env[62692]: DEBUG oslo_concurrency.lockutils [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 15.256s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.793432] env[62692]: DEBUG nova.network.neutron [-] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.823989] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c5eb8b2c-7f29-46cc-a478-b4cb8b73a811 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "interface-668fb0ea-7eed-4198-943c-8f916ec11368-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.076s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.928485] env[62692]: DEBUG oslo_vmware.api [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141624, 'name': PowerOffVM_Task, 'duration_secs': 0.242724} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.928775] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1015.928955] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1015.929227] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5bd505c8-4d58-49fd-b2da-992483abb92f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.990721] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1015.991106] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1015.991314] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Deleting the datastore file [datastore2] 9b8e830c-61b7-4dd2-8324-d3a96eec1465 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1015.991642] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-25b3d45b-b5db-428d-9904-e0f81d584242 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.998105] env[62692]: DEBUG oslo_vmware.api [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 1015.998105] env[62692]: value = "task-1141626" [ 1015.998105] env[62692]: _type = "Task" [ 1015.998105] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.006328] env[62692]: DEBUG oslo_vmware.api [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141626, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.107720] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4924eaa1-9d9c-484d-b19c-eff20b585bd5 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Lock "f523c482-8365-47b8-9398-a56a98fcffe8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.867s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.126254] env[62692]: DEBUG nova.compute.manager [req-f49b5144-1759-47c3-934b-5cf3243fb4a2 req-ab6d214b-43bf-477c-8a8d-bdea3e5b0f36 service nova] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Received event network-vif-deleted-2c7422d2-dfef-4010-a99a-60c510c54a3c {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1016.126467] env[62692]: DEBUG nova.compute.manager [req-f49b5144-1759-47c3-934b-5cf3243fb4a2 req-ab6d214b-43bf-477c-8a8d-bdea3e5b0f36 service nova] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Received event network-vif-deleted-dd324771-27d4-4505-b030-c8d8d9565968 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1016.126642] env[62692]: INFO nova.compute.manager [req-f49b5144-1759-47c3-934b-5cf3243fb4a2 req-ab6d214b-43bf-477c-8a8d-bdea3e5b0f36 service nova] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Neutron deleted interface dd324771-27d4-4505-b030-c8d8d9565968; detaching it from the instance and deleting it from the info cache [ 1016.127606] env[62692]: DEBUG nova.network.neutron [req-f49b5144-1759-47c3-934b-5cf3243fb4a2 req-ab6d214b-43bf-477c-8a8d-bdea3e5b0f36 service nova] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.147694] env[62692]: DEBUG nova.compute.manager [req-75b8b2a1-bed1-4438-b9e8-ded867ea265b req-97b01524-620d-4d26-b52f-09807cb67296 service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Received event network-vif-plugged-c43f9978-4cfd-467d-86e0-4c8acfefc821 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1016.147868] env[62692]: DEBUG oslo_concurrency.lockutils [req-75b8b2a1-bed1-4438-b9e8-ded867ea265b req-97b01524-620d-4d26-b52f-09807cb67296 service nova] Acquiring lock "1b753882-9118-479c-aa3e-cbb8bdf2e086-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.148305] env[62692]: DEBUG oslo_concurrency.lockutils [req-75b8b2a1-bed1-4438-b9e8-ded867ea265b req-97b01524-620d-4d26-b52f-09807cb67296 service nova] Lock "1b753882-9118-479c-aa3e-cbb8bdf2e086-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.148523] env[62692]: DEBUG oslo_concurrency.lockutils [req-75b8b2a1-bed1-4438-b9e8-ded867ea265b req-97b01524-620d-4d26-b52f-09807cb67296 service nova] Lock "1b753882-9118-479c-aa3e-cbb8bdf2e086-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.148710] env[62692]: DEBUG nova.compute.manager [req-75b8b2a1-bed1-4438-b9e8-ded867ea265b req-97b01524-620d-4d26-b52f-09807cb67296 service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] No waiting events found dispatching network-vif-plugged-c43f9978-4cfd-467d-86e0-4c8acfefc821 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1016.148905] env[62692]: WARNING nova.compute.manager [req-75b8b2a1-bed1-4438-b9e8-ded867ea265b req-97b01524-620d-4d26-b52f-09807cb67296 service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Received unexpected event network-vif-plugged-c43f9978-4cfd-467d-86e0-4c8acfefc821 for instance with vm_state building and task_state spawning. [ 1016.154799] env[62692]: DEBUG nova.network.neutron [-] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.183017] env[62692]: INFO nova.compute.manager [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Shelve offloading [ 1016.186208] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1016.186208] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-badb11bc-2442-45a8-b108-75b40b0de6a3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.192692] env[62692]: DEBUG oslo_vmware.api [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1016.192692] env[62692]: value = "task-1141627" [ 1016.192692] env[62692]: _type = "Task" [ 1016.192692] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.202500] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] VM already powered off {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1016.202699] env[62692]: DEBUG nova.compute.manager [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1016.203446] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-423cc8a8-dc03-4a04-b740-2d1f861e2ba6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.210180] env[62692]: DEBUG oslo_concurrency.lockutils [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "refresh_cache-250b9cc2-706e-4fbf-887c-1c80fc27a0ed" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.210394] env[62692]: DEBUG oslo_concurrency.lockutils [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired lock "refresh_cache-250b9cc2-706e-4fbf-887c-1c80fc27a0ed" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.210620] env[62692]: DEBUG nova.network.neutron [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1016.238627] env[62692]: DEBUG nova.network.neutron [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Successfully updated port: c43f9978-4cfd-467d-86e0-4c8acfefc821 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1016.301830] env[62692]: INFO nova.compute.manager [-] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Took 1.49 seconds to deallocate network for instance. [ 1016.302329] env[62692]: DEBUG oslo_concurrency.lockutils [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "refresh_cache-d99737d8-2eb0-40ee-b61e-6c736c84ea59" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.302532] env[62692]: DEBUG oslo_concurrency.lockutils [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired lock "refresh_cache-d99737d8-2eb0-40ee-b61e-6c736c84ea59" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.302721] env[62692]: DEBUG nova.network.neutron [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1016.456585] env[62692]: DEBUG nova.compute.manager [None req-aaa4585e-bc34-4642-8819-493fd6cf5383 tempest-ServerExternalEventsTest-1302334229 tempest-ServerExternalEventsTest-1302334229-project] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Received event network-changed {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1016.456847] env[62692]: DEBUG nova.compute.manager [None req-aaa4585e-bc34-4642-8819-493fd6cf5383 tempest-ServerExternalEventsTest-1302334229 tempest-ServerExternalEventsTest-1302334229-project] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Refreshing instance network info cache due to event network-changed. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1016.457500] env[62692]: DEBUG oslo_concurrency.lockutils [None req-aaa4585e-bc34-4642-8819-493fd6cf5383 tempest-ServerExternalEventsTest-1302334229 tempest-ServerExternalEventsTest-1302334229-project] Acquiring lock "refresh_cache-f523c482-8365-47b8-9398-a56a98fcffe8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.457694] env[62692]: DEBUG oslo_concurrency.lockutils [None req-aaa4585e-bc34-4642-8819-493fd6cf5383 tempest-ServerExternalEventsTest-1302334229 tempest-ServerExternalEventsTest-1302334229-project] Acquired lock "refresh_cache-f523c482-8365-47b8-9398-a56a98fcffe8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.457895] env[62692]: DEBUG nova.network.neutron [None req-aaa4585e-bc34-4642-8819-493fd6cf5383 tempest-ServerExternalEventsTest-1302334229 tempest-ServerExternalEventsTest-1302334229-project] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1016.515109] env[62692]: DEBUG oslo_vmware.api [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141626, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.352307} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.515565] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1016.515891] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1016.516220] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1016.516546] env[62692]: INFO nova.compute.manager [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1016.516937] env[62692]: DEBUG oslo.service.loopingcall [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1016.517261] env[62692]: DEBUG nova.compute.manager [-] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1016.517542] env[62692]: DEBUG nova.network.neutron [-] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1016.563326] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f585a8e8-f646-412a-8df0-708df11af9ba {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.571399] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1e93bfb-b3f4-4611-9b13-d69cda2555d9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.605949] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a1fe091-bab4-4e61-a916-2412b1868175 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.615114] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e5c2511-fcba-4f59-a041-ff37e4091fe1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.628877] env[62692]: DEBUG nova.compute.provider_tree [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1016.631179] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-50aa5155-1221-48e8-90e0-15b29b247a65 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.640069] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77cf57df-7da0-4ced-a80a-e93ae41219c5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.658394] env[62692]: INFO nova.compute.manager [-] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Took 1.26 seconds to deallocate network for instance. [ 1016.676521] env[62692]: DEBUG nova.compute.manager [req-f49b5144-1759-47c3-934b-5cf3243fb4a2 req-ab6d214b-43bf-477c-8a8d-bdea3e5b0f36 service nova] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Detach interface failed, port_id=dd324771-27d4-4505-b030-c8d8d9565968, reason: Instance f6085887-a5d3-4096-be38-d843bb54304a could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1016.742388] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquiring lock "refresh_cache-1b753882-9118-479c-aa3e-cbb8bdf2e086" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.742388] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquired lock "refresh_cache-1b753882-9118-479c-aa3e-cbb8bdf2e086" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.742388] env[62692]: DEBUG nova.network.neutron [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1016.810312] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.048438] env[62692]: DEBUG nova.network.neutron [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Updating instance_info_cache with network_info: [{"id": "30d6694e-3be7-4c78-96c8-c9e2005856a5", "address": "fa:16:3e:b5:ab:9f", "network": {"id": "dd260020-2d74-46ba-aa78-af29bdefa176", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-617179655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "589895ae20394e2fa525707bf7d09f2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30d6694e-3b", "ovs_interfaceid": "30d6694e-3be7-4c78-96c8-c9e2005856a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.136822] env[62692]: DEBUG nova.network.neutron [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Updating instance_info_cache with network_info: [{"id": "7f2cabdf-cbda-4cf7-a518-13f619716d12", "address": "fa:16:3e:6b:22:0c", "network": {"id": "f70727d6-a9aa-44b0-a244-2533b3f96571", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-284105837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1037e5d436bd429391159b87a712f6e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f2cabdf-cb", "ovs_interfaceid": "7f2cabdf-cbda-4cf7-a518-13f619716d12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.157589] env[62692]: ERROR nova.scheduler.client.report [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [req-a038555f-0258-4a8e-acd4-ff74cd484f4a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 470758a8-5efa-4875-aad5-f512a727752f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-a038555f-0258-4a8e-acd4-ff74cd484f4a"}]} [ 1017.165432] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.165862] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Acquiring lock "f523c482-8365-47b8-9398-a56a98fcffe8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.166027] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Lock "f523c482-8365-47b8-9398-a56a98fcffe8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.166219] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Acquiring lock "f523c482-8365-47b8-9398-a56a98fcffe8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.166460] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Lock "f523c482-8365-47b8-9398-a56a98fcffe8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.167075] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Lock "f523c482-8365-47b8-9398-a56a98fcffe8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.170891] env[62692]: INFO nova.compute.manager [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Terminating instance [ 1017.171243] env[62692]: DEBUG nova.compute.manager [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1017.171440] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1017.172313] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8865773-7653-41eb-b714-f7c431010a2e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.178575] env[62692]: DEBUG nova.scheduler.client.report [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Refreshing inventories for resource provider 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1017.183968] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1017.186488] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4e4b84d4-4d63-4c46-8289-146b80c792b9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.194577] env[62692]: DEBUG oslo_vmware.api [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Waiting for the task: (returnval){ [ 1017.194577] env[62692]: value = "task-1141628" [ 1017.194577] env[62692]: _type = "Task" [ 1017.194577] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.199277] env[62692]: DEBUG nova.scheduler.client.report [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Updating ProviderTree inventory for provider 470758a8-5efa-4875-aad5-f512a727752f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1017.199813] env[62692]: DEBUG nova.compute.provider_tree [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 165, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1017.207881] env[62692]: DEBUG oslo_vmware.api [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Task: {'id': task-1141628, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.213815] env[62692]: DEBUG nova.scheduler.client.report [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Refreshing aggregate associations for resource provider 470758a8-5efa-4875-aad5-f512a727752f, aggregates: None {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1017.238672] env[62692]: DEBUG nova.scheduler.client.report [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Refreshing trait associations for resource provider 470758a8-5efa-4875-aad5-f512a727752f, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1017.281589] env[62692]: DEBUG nova.network.neutron [None req-aaa4585e-bc34-4642-8819-493fd6cf5383 tempest-ServerExternalEventsTest-1302334229 tempest-ServerExternalEventsTest-1302334229-project] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Updating instance_info_cache with network_info: [{"id": "9867c68f-a0ff-499d-9b91-47cbd25c4b69", "address": "fa:16:3e:9b:20:fd", "network": {"id": "a74c4d30-3568-4e50-98f5-d2fcc05ec728", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.220", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "989968e8796a4a70a4dbd045418a475f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9867c68f-a0", "ovs_interfaceid": "9867c68f-a0ff-499d-9b91-47cbd25c4b69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.302718] env[62692]: DEBUG nova.network.neutron [-] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.310568] env[62692]: DEBUG nova.network.neutron [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1017.526160] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12f25d65-c4da-4615-a8ca-82eff9cc6445 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.532587] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea6dd5c9-c08f-4a9e-ad7f-f2ea1d098968 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.565025] env[62692]: DEBUG oslo_concurrency.lockutils [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Releasing lock "refresh_cache-250b9cc2-706e-4fbf-887c-1c80fc27a0ed" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.565997] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb45ac3-3488-46b9-8b65-4b1bf20480e6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.571122] env[62692]: DEBUG nova.network.neutron [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Updating instance_info_cache with network_info: [{"id": "c43f9978-4cfd-467d-86e0-4c8acfefc821", "address": "fa:16:3e:ef:d0:33", "network": {"id": "b7c23ada-5498-4686-bd7b-b32d0837e401", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1234309205-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f26b93a4a43b45e8a11f3793996c8496", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b67e519-46cf-44ce-b670-4ba4c0c5b658", "external-id": "nsx-vlan-transportzone-110", "segmentation_id": 110, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc43f9978-4c", "ovs_interfaceid": "c43f9978-4cfd-467d-86e0-4c8acfefc821", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.576116] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b893bba-2694-4c44-9163-86e67079c259 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.592398] env[62692]: DEBUG nova.compute.provider_tree [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1017.642331] env[62692]: DEBUG oslo_concurrency.lockutils [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Releasing lock "refresh_cache-d99737d8-2eb0-40ee-b61e-6c736c84ea59" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.711422] env[62692]: DEBUG oslo_vmware.api [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Task: {'id': task-1141628, 'name': PowerOffVM_Task, 'duration_secs': 0.192889} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.714426] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1017.714426] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1017.714426] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-22c75ea0-a5c3-4e4f-a63a-0fe1b97ee7b4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.782929] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1017.783204] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1017.783373] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Deleting the datastore file [datastore2] f523c482-8365-47b8-9398-a56a98fcffe8 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1017.783642] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-07cd75f6-775a-4cb6-8d2a-075b9f20eb37 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.785903] env[62692]: DEBUG oslo_concurrency.lockutils [None req-aaa4585e-bc34-4642-8819-493fd6cf5383 tempest-ServerExternalEventsTest-1302334229 tempest-ServerExternalEventsTest-1302334229-project] Releasing lock "refresh_cache-f523c482-8365-47b8-9398-a56a98fcffe8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.792494] env[62692]: DEBUG oslo_vmware.api [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Waiting for the task: (returnval){ [ 1017.792494] env[62692]: value = "task-1141630" [ 1017.792494] env[62692]: _type = "Task" [ 1017.792494] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.801307] env[62692]: DEBUG oslo_vmware.api [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Task: {'id': task-1141630, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.803849] env[62692]: INFO nova.compute.manager [-] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Took 1.29 seconds to deallocate network for instance. [ 1017.863949] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1017.865139] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-479a58cb-7a91-480c-bead-faeb07f66aca {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.873210] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1017.873478] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2cc0e51b-7584-461d-bb13-3771a3c5d6c6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.932032] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1017.932305] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1017.932487] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Deleting the datastore file [datastore1] 250b9cc2-706e-4fbf-887c-1c80fc27a0ed {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1017.932760] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9b5baa48-4019-463a-a486-8ac6556fc456 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.939300] env[62692]: DEBUG oslo_vmware.api [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1017.939300] env[62692]: value = "task-1141632" [ 1017.939300] env[62692]: _type = "Task" [ 1017.939300] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.956932] env[62692]: DEBUG oslo_vmware.api [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141632, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.072430] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Releasing lock "refresh_cache-1b753882-9118-479c-aa3e-cbb8bdf2e086" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.072834] env[62692]: DEBUG nova.compute.manager [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Instance network_info: |[{"id": "c43f9978-4cfd-467d-86e0-4c8acfefc821", "address": "fa:16:3e:ef:d0:33", "network": {"id": "b7c23ada-5498-4686-bd7b-b32d0837e401", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1234309205-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f26b93a4a43b45e8a11f3793996c8496", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b67e519-46cf-44ce-b670-4ba4c0c5b658", "external-id": "nsx-vlan-transportzone-110", "segmentation_id": 110, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc43f9978-4c", "ovs_interfaceid": "c43f9978-4cfd-467d-86e0-4c8acfefc821", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1018.073303] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ef:d0:33', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3b67e519-46cf-44ce-b670-4ba4c0c5b658', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c43f9978-4cfd-467d-86e0-4c8acfefc821', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1018.081294] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Creating folder: Project (f26b93a4a43b45e8a11f3793996c8496). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1018.081294] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-25a98303-cbbf-4410-a300-7a97861c8342 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.092623] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Created folder: Project (f26b93a4a43b45e8a11f3793996c8496) in parent group-v248868. [ 1018.092838] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Creating folder: Instances. Parent ref: group-v249023. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1018.095597] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5a2e821d-aa12-49e6-8180-087907e6eff9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.104988] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Created folder: Instances in parent group-v249023. [ 1018.105294] env[62692]: DEBUG oslo.service.loopingcall [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1018.105519] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1018.105767] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-02068a10-66fd-4188-9022-e4a55db47f1a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.130055] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1018.130055] env[62692]: value = "task-1141635" [ 1018.130055] env[62692]: _type = "Task" [ 1018.130055] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.130055] env[62692]: DEBUG nova.scheduler.client.report [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Updated inventory for provider 470758a8-5efa-4875-aad5-f512a727752f with generation 113 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1018.130055] env[62692]: DEBUG nova.compute.provider_tree [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Updating resource provider 470758a8-5efa-4875-aad5-f512a727752f generation from 113 to 114 during operation: update_inventory {{(pid=62692) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1018.130055] env[62692]: DEBUG nova.compute.provider_tree [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1018.143818] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141635, 'name': CreateVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.195740] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "interface-668fb0ea-7eed-4198-943c-8f916ec11368-dac9e9e1-d8ad-4ee7-8836-a84a7cb3cf27" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.195740] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "interface-668fb0ea-7eed-4198-943c-8f916ec11368-dac9e9e1-d8ad-4ee7-8836-a84a7cb3cf27" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.196535] env[62692]: DEBUG nova.objects.instance [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lazy-loading 'flavor' on Instance uuid 668fb0ea-7eed-4198-943c-8f916ec11368 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.216155] env[62692]: DEBUG nova.compute.manager [req-b4f9ee23-566d-48bb-a2f7-1a024eaff474 req-764ef155-81ca-4721-b9a3-be6b5f2c31da service nova] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Received event network-vif-unplugged-30d6694e-3be7-4c78-96c8-c9e2005856a5 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1018.216390] env[62692]: DEBUG oslo_concurrency.lockutils [req-b4f9ee23-566d-48bb-a2f7-1a024eaff474 req-764ef155-81ca-4721-b9a3-be6b5f2c31da service nova] Acquiring lock "250b9cc2-706e-4fbf-887c-1c80fc27a0ed-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.216611] env[62692]: DEBUG oslo_concurrency.lockutils [req-b4f9ee23-566d-48bb-a2f7-1a024eaff474 req-764ef155-81ca-4721-b9a3-be6b5f2c31da service nova] Lock "250b9cc2-706e-4fbf-887c-1c80fc27a0ed-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.216809] env[62692]: DEBUG oslo_concurrency.lockutils [req-b4f9ee23-566d-48bb-a2f7-1a024eaff474 req-764ef155-81ca-4721-b9a3-be6b5f2c31da service nova] Lock "250b9cc2-706e-4fbf-887c-1c80fc27a0ed-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.216986] env[62692]: DEBUG nova.compute.manager [req-b4f9ee23-566d-48bb-a2f7-1a024eaff474 req-764ef155-81ca-4721-b9a3-be6b5f2c31da service nova] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] No waiting events found dispatching network-vif-unplugged-30d6694e-3be7-4c78-96c8-c9e2005856a5 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1018.217176] env[62692]: WARNING nova.compute.manager [req-b4f9ee23-566d-48bb-a2f7-1a024eaff474 req-764ef155-81ca-4721-b9a3-be6b5f2c31da service nova] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Received unexpected event network-vif-unplugged-30d6694e-3be7-4c78-96c8-c9e2005856a5 for instance with vm_state shelved and task_state shelving_offloading. [ 1018.217451] env[62692]: DEBUG nova.compute.manager [req-b4f9ee23-566d-48bb-a2f7-1a024eaff474 req-764ef155-81ca-4721-b9a3-be6b5f2c31da service nova] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Received event network-changed-30d6694e-3be7-4c78-96c8-c9e2005856a5 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1018.217525] env[62692]: DEBUG nova.compute.manager [req-b4f9ee23-566d-48bb-a2f7-1a024eaff474 req-764ef155-81ca-4721-b9a3-be6b5f2c31da service nova] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Refreshing instance network info cache due to event network-changed-30d6694e-3be7-4c78-96c8-c9e2005856a5. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1018.217732] env[62692]: DEBUG oslo_concurrency.lockutils [req-b4f9ee23-566d-48bb-a2f7-1a024eaff474 req-764ef155-81ca-4721-b9a3-be6b5f2c31da service nova] Acquiring lock "refresh_cache-250b9cc2-706e-4fbf-887c-1c80fc27a0ed" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.217939] env[62692]: DEBUG oslo_concurrency.lockutils [req-b4f9ee23-566d-48bb-a2f7-1a024eaff474 req-764ef155-81ca-4721-b9a3-be6b5f2c31da service nova] Acquired lock "refresh_cache-250b9cc2-706e-4fbf-887c-1c80fc27a0ed" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.218046] env[62692]: DEBUG nova.network.neutron [req-b4f9ee23-566d-48bb-a2f7-1a024eaff474 req-764ef155-81ca-4721-b9a3-be6b5f2c31da service nova] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Refreshing network info cache for port 30d6694e-3be7-4c78-96c8-c9e2005856a5 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1018.239284] env[62692]: DEBUG nova.compute.manager [req-a6597ced-bca8-4771-8a6d-26792c14cd3c req-91facd72-70ad-49c7-a4c8-b00452f4ebb7 service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Received event network-changed-c43f9978-4cfd-467d-86e0-4c8acfefc821 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1018.239509] env[62692]: DEBUG nova.compute.manager [req-a6597ced-bca8-4771-8a6d-26792c14cd3c req-91facd72-70ad-49c7-a4c8-b00452f4ebb7 service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Refreshing instance network info cache due to event network-changed-c43f9978-4cfd-467d-86e0-4c8acfefc821. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1018.239577] env[62692]: DEBUG oslo_concurrency.lockutils [req-a6597ced-bca8-4771-8a6d-26792c14cd3c req-91facd72-70ad-49c7-a4c8-b00452f4ebb7 service nova] Acquiring lock "refresh_cache-1b753882-9118-479c-aa3e-cbb8bdf2e086" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.239733] env[62692]: DEBUG oslo_concurrency.lockutils [req-a6597ced-bca8-4771-8a6d-26792c14cd3c req-91facd72-70ad-49c7-a4c8-b00452f4ebb7 service nova] Acquired lock "refresh_cache-1b753882-9118-479c-aa3e-cbb8bdf2e086" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.239865] env[62692]: DEBUG nova.network.neutron [req-a6597ced-bca8-4771-8a6d-26792c14cd3c req-91facd72-70ad-49c7-a4c8-b00452f4ebb7 service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Refreshing network info cache for port c43f9978-4cfd-467d-86e0-4c8acfefc821 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1018.301708] env[62692]: DEBUG oslo_vmware.api [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Task: {'id': task-1141630, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.178028} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.301985] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1018.302190] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1018.302368] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1018.302543] env[62692]: INFO nova.compute.manager [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1018.302786] env[62692]: DEBUG oslo.service.loopingcall [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1018.302978] env[62692]: DEBUG nova.compute.manager [-] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1018.303087] env[62692]: DEBUG nova.network.neutron [-] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1018.310459] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.451345] env[62692]: DEBUG oslo_vmware.api [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141632, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.2288} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.451650] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1018.451946] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1018.452229] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1018.473742] env[62692]: INFO nova.scheduler.client.report [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Deleted allocations for instance 250b9cc2-706e-4fbf-887c-1c80fc27a0ed [ 1019.267469] env[62692]: DEBUG oslo_concurrency.lockutils [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.267751] env[62692]: DEBUG nova.network.neutron [-] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.286371] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141635, 'name': CreateVM_Task, 'duration_secs': 0.400746} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.286839] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1019.287683] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.287853] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.288192] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1019.288826] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25badc8c-11cd-41e8-9d82-0a1847c4dd0e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.298042] env[62692]: DEBUG oslo_vmware.api [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Waiting for the task: (returnval){ [ 1019.298042] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]522ba79c-56a7-f081-4cb2-0f469f058db9" [ 1019.298042] env[62692]: _type = "Task" [ 1019.298042] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.310231] env[62692]: DEBUG oslo_vmware.api [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]522ba79c-56a7-f081-4cb2-0f469f058db9, 'name': SearchDatastore_Task, 'duration_secs': 0.010147} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.310231] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.310231] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1019.310231] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.310231] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.310231] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1019.310611] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eb22f783-a789-430a-b3ed-931abb199d71 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.318811] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1019.318983] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1019.319705] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a8df5a9-ac86-4d10-b3d2-4bd093e962bf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.324892] env[62692]: DEBUG oslo_vmware.api [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Waiting for the task: (returnval){ [ 1019.324892] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]524f3b4b-4774-c49c-22c6-5c681014cc94" [ 1019.324892] env[62692]: _type = "Task" [ 1019.324892] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.332278] env[62692]: DEBUG oslo_vmware.api [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]524f3b4b-4774-c49c-22c6-5c681014cc94, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.564088] env[62692]: DEBUG nova.network.neutron [req-b4f9ee23-566d-48bb-a2f7-1a024eaff474 req-764ef155-81ca-4721-b9a3-be6b5f2c31da service nova] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Updated VIF entry in instance network info cache for port 30d6694e-3be7-4c78-96c8-c9e2005856a5. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1019.564524] env[62692]: DEBUG nova.network.neutron [req-b4f9ee23-566d-48bb-a2f7-1a024eaff474 req-764ef155-81ca-4721-b9a3-be6b5f2c31da service nova] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Updating instance_info_cache with network_info: [{"id": "30d6694e-3be7-4c78-96c8-c9e2005856a5", "address": "fa:16:3e:b5:ab:9f", "network": {"id": "dd260020-2d74-46ba-aa78-af29bdefa176", "bridge": null, "label": "tempest-DeleteServersTestJSON-617179655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "589895ae20394e2fa525707bf7d09f2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap30d6694e-3b", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.659235] env[62692]: DEBUG nova.objects.instance [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lazy-loading 'pci_requests' on Instance uuid 668fb0ea-7eed-4198-943c-8f916ec11368 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1019.752040] env[62692]: DEBUG nova.network.neutron [req-a6597ced-bca8-4771-8a6d-26792c14cd3c req-91facd72-70ad-49c7-a4c8-b00452f4ebb7 service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Updated VIF entry in instance network info cache for port c43f9978-4cfd-467d-86e0-4c8acfefc821. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1019.752040] env[62692]: DEBUG nova.network.neutron [req-a6597ced-bca8-4771-8a6d-26792c14cd3c req-91facd72-70ad-49c7-a4c8-b00452f4ebb7 service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Updating instance_info_cache with network_info: [{"id": "c43f9978-4cfd-467d-86e0-4c8acfefc821", "address": "fa:16:3e:ef:d0:33", "network": {"id": "b7c23ada-5498-4686-bd7b-b32d0837e401", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1234309205-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f26b93a4a43b45e8a11f3793996c8496", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b67e519-46cf-44ce-b670-4ba4c0c5b658", "external-id": "nsx-vlan-transportzone-110", "segmentation_id": 110, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc43f9978-4c", "ovs_interfaceid": "c43f9978-4cfd-467d-86e0-4c8acfefc821", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.773562] env[62692]: DEBUG oslo_concurrency.lockutils [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 3.981s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.778340] env[62692]: INFO nova.compute.manager [-] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Took 1.48 seconds to deallocate network for instance. [ 1019.778676] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.436s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.780211] env[62692]: INFO nova.compute.claims [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1019.785330] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36d44585-570c-4fdf-8ae1-70c026bac276 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.792727] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Acquiring lock "16790505-cdf6-4937-8839-fa685a5f413e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.793074] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Lock "16790505-cdf6-4937-8839-fa685a5f413e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.793172] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Acquiring lock "16790505-cdf6-4937-8839-fa685a5f413e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.793383] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Lock "16790505-cdf6-4937-8839-fa685a5f413e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.793634] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Lock "16790505-cdf6-4937-8839-fa685a5f413e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.809386] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Updating instance 'd99737d8-2eb0-40ee-b61e-6c736c84ea59' progress to 0 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1019.813463] env[62692]: INFO nova.compute.manager [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Terminating instance [ 1019.815468] env[62692]: DEBUG nova.compute.manager [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1019.815713] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1019.816609] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9a6127d-1fe1-4ae9-8fae-e719cd354275 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.824424] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1019.824693] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8d349c7f-e9c1-479a-b59e-5c71ba414a78 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.837630] env[62692]: DEBUG oslo_vmware.api [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]524f3b4b-4774-c49c-22c6-5c681014cc94, 'name': SearchDatastore_Task, 'duration_secs': 0.008528} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.839504] env[62692]: DEBUG oslo_vmware.api [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Waiting for the task: (returnval){ [ 1019.839504] env[62692]: value = "task-1141636" [ 1019.839504] env[62692]: _type = "Task" [ 1019.839504] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.839714] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75cbf834-5b96-4131-bdec-067f70968c09 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.851293] env[62692]: DEBUG oslo_vmware.api [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Task: {'id': task-1141636, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.852760] env[62692]: DEBUG oslo_vmware.api [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Waiting for the task: (returnval){ [ 1019.852760] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]526f9560-16b8-b5c4-bc75-32c64faf46d7" [ 1019.852760] env[62692]: _type = "Task" [ 1019.852760] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.864408] env[62692]: DEBUG oslo_vmware.api [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]526f9560-16b8-b5c4-bc75-32c64faf46d7, 'name': SearchDatastore_Task, 'duration_secs': 0.0108} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.864568] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.865273] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 1b753882-9118-479c-aa3e-cbb8bdf2e086/1b753882-9118-479c-aa3e-cbb8bdf2e086.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1019.865273] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7e422402-9759-41c5-a76a-b257d053d8a9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.871916] env[62692]: DEBUG oslo_vmware.api [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Waiting for the task: (returnval){ [ 1019.871916] env[62692]: value = "task-1141637" [ 1019.871916] env[62692]: _type = "Task" [ 1019.871916] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.879805] env[62692]: DEBUG oslo_vmware.api [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141637, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.068196] env[62692]: DEBUG oslo_concurrency.lockutils [req-b4f9ee23-566d-48bb-a2f7-1a024eaff474 req-764ef155-81ca-4721-b9a3-be6b5f2c31da service nova] Releasing lock "refresh_cache-250b9cc2-706e-4fbf-887c-1c80fc27a0ed" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.163235] env[62692]: DEBUG nova.objects.base [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Object Instance<668fb0ea-7eed-4198-943c-8f916ec11368> lazy-loaded attributes: flavor,pci_requests {{(pid=62692) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1020.163235] env[62692]: DEBUG nova.network.neutron [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1020.251656] env[62692]: DEBUG nova.policy [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a251ef5f6cb143178928babc75a0fef0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ba97a775e99a463c845518b0df8cd3e2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1020.254526] env[62692]: DEBUG oslo_concurrency.lockutils [req-a6597ced-bca8-4771-8a6d-26792c14cd3c req-91facd72-70ad-49c7-a4c8-b00452f4ebb7 service nova] Releasing lock "refresh_cache-1b753882-9118-479c-aa3e-cbb8bdf2e086" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.254823] env[62692]: DEBUG nova.compute.manager [req-a6597ced-bca8-4771-8a6d-26792c14cd3c req-91facd72-70ad-49c7-a4c8-b00452f4ebb7 service nova] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Received event network-vif-deleted-cad44f99-9f63-4d6e-a084-4061f88e126f {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1020.295364] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.317810] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1020.318117] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-890dc065-f3e7-4c15-831d-6a870ef79384 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.329087] env[62692]: DEBUG oslo_vmware.api [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1020.329087] env[62692]: value = "task-1141638" [ 1020.329087] env[62692]: _type = "Task" [ 1020.329087] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.338992] env[62692]: DEBUG oslo_vmware.api [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141638, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.354059] env[62692]: DEBUG oslo_vmware.api [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Task: {'id': task-1141636, 'name': PowerOffVM_Task, 'duration_secs': 0.192499} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.354643] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1020.355308] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1020.355308] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ef60a849-96b2-4db5-9cf2-90d045ae65cb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.359644] env[62692]: INFO nova.scheduler.client.report [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Deleted allocation for migration 4c7b8a13-dd0a-48d9-b1fe-523c53945e0c [ 1020.385136] env[62692]: DEBUG oslo_vmware.api [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141637, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.497929} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.385524] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 1b753882-9118-479c-aa3e-cbb8bdf2e086/1b753882-9118-479c-aa3e-cbb8bdf2e086.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1020.385842] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1020.386207] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2c4ab121-9d57-4bbe-a483-960742f7c459 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.394895] env[62692]: DEBUG oslo_vmware.api [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Waiting for the task: (returnval){ [ 1020.394895] env[62692]: value = "task-1141640" [ 1020.394895] env[62692]: _type = "Task" [ 1020.394895] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.404303] env[62692]: DEBUG oslo_vmware.api [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141640, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.429349] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1020.429694] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1020.429934] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Deleting the datastore file [datastore1] 16790505-cdf6-4937-8839-fa685a5f413e {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1020.430247] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-80abb160-2429-47db-bbf7-824f8c53bfdc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.437040] env[62692]: DEBUG oslo_vmware.api [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Waiting for the task: (returnval){ [ 1020.437040] env[62692]: value = "task-1141641" [ 1020.437040] env[62692]: _type = "Task" [ 1020.437040] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.448098] env[62692]: DEBUG oslo_vmware.api [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Task: {'id': task-1141641, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.492773] env[62692]: DEBUG oslo_concurrency.lockutils [None req-14d6ffc0-8145-4ba8-9f78-a28d05698ae8 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "250b9cc2-706e-4fbf-887c-1c80fc27a0ed" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.555067] env[62692]: DEBUG nova.compute.manager [req-ca1edf8e-7eef-4276-8f8d-8426b5679742 req-b14c9fba-3818-4955-8435-1741d4b0885f service nova] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Received event network-vif-deleted-9867c68f-a0ff-499d-9b91-47cbd25c4b69 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1020.841451] env[62692]: DEBUG oslo_vmware.api [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141638, 'name': PowerOffVM_Task, 'duration_secs': 0.196939} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.841646] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1020.841861] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Updating instance 'd99737d8-2eb0-40ee-b61e-6c736c84ea59' progress to 17 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1020.868634] env[62692]: DEBUG oslo_concurrency.lockutils [None req-84ad73be-1190-406c-ab21-92e2a8fcff95 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "3f6c9744-a6e3-43f3-8b6b-624ba178747d" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 24.029s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.906711] env[62692]: DEBUG oslo_vmware.api [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141640, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067466} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.906711] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1020.906711] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4df461ec-fe78-46fa-a0d6-1a865334a4f8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.929289] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] 1b753882-9118-479c-aa3e-cbb8bdf2e086/1b753882-9118-479c-aa3e-cbb8bdf2e086.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1020.932464] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4ebf3e05-96e4-478c-97e7-2424431a9990 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.956254] env[62692]: DEBUG oslo_vmware.api [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Task: {'id': task-1141641, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.186553} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.959802] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1020.960016] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1020.960213] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1020.960401] env[62692]: INFO nova.compute.manager [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1020.960707] env[62692]: DEBUG oslo.service.loopingcall [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1020.960994] env[62692]: DEBUG oslo_vmware.api [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Waiting for the task: (returnval){ [ 1020.960994] env[62692]: value = "task-1141642" [ 1020.960994] env[62692]: _type = "Task" [ 1020.960994] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.961386] env[62692]: DEBUG nova.compute.manager [-] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1020.961485] env[62692]: DEBUG nova.network.neutron [-] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1020.971222] env[62692]: DEBUG oslo_vmware.api [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141642, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.083274] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f15df60d-49e8-4bc1-96e2-2d0c591541db {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.091015] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8ebeab6-a0cb-4c58-8b14-4e9323a67918 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.120825] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebb21c5b-eeae-4c30-bd35-e717c1e939dc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.128295] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d71b20ad-0aa5-4bdf-81ee-2c56918ca9d3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.145175] env[62692]: DEBUG nova.compute.provider_tree [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1021.350249] env[62692]: DEBUG nova.virt.hardware [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1021.350609] env[62692]: DEBUG nova.virt.hardware [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1021.350752] env[62692]: DEBUG nova.virt.hardware [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1021.350936] env[62692]: DEBUG nova.virt.hardware [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1021.351110] env[62692]: DEBUG nova.virt.hardware [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1021.351264] env[62692]: DEBUG nova.virt.hardware [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1021.351471] env[62692]: DEBUG nova.virt.hardware [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1021.351637] env[62692]: DEBUG nova.virt.hardware [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1021.351806] env[62692]: DEBUG nova.virt.hardware [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1021.351972] env[62692]: DEBUG nova.virt.hardware [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1021.352163] env[62692]: DEBUG nova.virt.hardware [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1021.357262] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f584d6fe-16af-4683-9253-06f257095471 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.374170] env[62692]: DEBUG oslo_vmware.api [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1021.374170] env[62692]: value = "task-1141643" [ 1021.374170] env[62692]: _type = "Task" [ 1021.374170] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.382591] env[62692]: DEBUG oslo_vmware.api [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141643, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.472866] env[62692]: DEBUG oslo_vmware.api [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141642, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.648925] env[62692]: DEBUG nova.scheduler.client.report [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1021.806264] env[62692]: DEBUG nova.network.neutron [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Successfully updated port: dac9e9e1-d8ad-4ee7-8836-a84a7cb3cf27 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1021.884705] env[62692]: DEBUG oslo_vmware.api [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141643, 'name': ReconfigVM_Task, 'duration_secs': 0.301714} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.885032] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Updating instance 'd99737d8-2eb0-40ee-b61e-6c736c84ea59' progress to 33 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1021.918725] env[62692]: DEBUG nova.network.neutron [-] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.974837] env[62692]: DEBUG oslo_vmware.api [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141642, 'name': ReconfigVM_Task, 'duration_secs': 0.707995} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.975173] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Reconfigured VM instance instance-00000058 to attach disk [datastore2] 1b753882-9118-479c-aa3e-cbb8bdf2e086/1b753882-9118-479c-aa3e-cbb8bdf2e086.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1021.979133] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d8dd283d-5434-48e2-8965-b8814e20d382 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.983420] env[62692]: DEBUG oslo_vmware.api [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Waiting for the task: (returnval){ [ 1021.983420] env[62692]: value = "task-1141644" [ 1021.983420] env[62692]: _type = "Task" [ 1021.983420] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.997794] env[62692]: DEBUG oslo_vmware.api [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141644, 'name': Rename_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.152779] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.374s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.153500] env[62692]: DEBUG nova.compute.manager [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1022.156728] env[62692]: DEBUG oslo_concurrency.lockutils [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.475s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.156830] env[62692]: DEBUG nova.objects.instance [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lazy-loading 'resources' on Instance uuid b23d3484-4fc5-46b2-8e8c-bdaab087f046 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1022.310123] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "refresh_cache-668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.310321] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "refresh_cache-668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.310517] env[62692]: DEBUG nova.network.neutron [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1022.394556] env[62692]: DEBUG nova.virt.hardware [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1022.394876] env[62692]: DEBUG nova.virt.hardware [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1022.395072] env[62692]: DEBUG nova.virt.hardware [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1022.395315] env[62692]: DEBUG nova.virt.hardware [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1022.395505] env[62692]: DEBUG nova.virt.hardware [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1022.395695] env[62692]: DEBUG nova.virt.hardware [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1022.396070] env[62692]: DEBUG nova.virt.hardware [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1022.396290] env[62692]: DEBUG nova.virt.hardware [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1022.396507] env[62692]: DEBUG nova.virt.hardware [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1022.396774] env[62692]: DEBUG nova.virt.hardware [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1022.396984] env[62692]: DEBUG nova.virt.hardware [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1022.403416] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Reconfiguring VM instance instance-00000021 to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1022.404016] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d68284b9-1f42-43e0-8a46-39bc0e69ba5c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.420636] env[62692]: INFO nova.compute.manager [-] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Took 1.46 seconds to deallocate network for instance. [ 1022.423629] env[62692]: DEBUG oslo_vmware.api [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1022.423629] env[62692]: value = "task-1141645" [ 1022.423629] env[62692]: _type = "Task" [ 1022.423629] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.436182] env[62692]: DEBUG oslo_vmware.api [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141645, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.493796] env[62692]: DEBUG oslo_vmware.api [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141644, 'name': Rename_Task, 'duration_secs': 0.141281} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.494091] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1022.494337] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c8cc3941-9c3d-4246-a9ac-981d8f8c2132 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.501009] env[62692]: DEBUG oslo_vmware.api [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Waiting for the task: (returnval){ [ 1022.501009] env[62692]: value = "task-1141646" [ 1022.501009] env[62692]: _type = "Task" [ 1022.501009] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.514249] env[62692]: DEBUG oslo_vmware.api [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141646, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.660043] env[62692]: DEBUG nova.compute.utils [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1022.660972] env[62692]: DEBUG nova.compute.manager [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1022.661161] env[62692]: DEBUG nova.network.neutron [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1022.705293] env[62692]: DEBUG nova.policy [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '016134ac36fd4bd893226ae95a955e28', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37921f67664544478cb2568def068b11', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1022.834759] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "3f6c9744-a6e3-43f3-8b6b-624ba178747d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.834860] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "3f6c9744-a6e3-43f3-8b6b-624ba178747d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.835524] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "3f6c9744-a6e3-43f3-8b6b-624ba178747d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.835524] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "3f6c9744-a6e3-43f3-8b6b-624ba178747d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.835524] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "3f6c9744-a6e3-43f3-8b6b-624ba178747d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.837473] env[62692]: INFO nova.compute.manager [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Terminating instance [ 1022.842819] env[62692]: DEBUG nova.compute.manager [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1022.843036] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1022.843883] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7765b15d-ceb2-47a5-8e0e-e23fb12c56bc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.851514] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1022.854249] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8401beec-9662-4e0b-ba1e-148d4bd4fbe1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.862215] env[62692]: DEBUG oslo_vmware.api [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1022.862215] env[62692]: value = "task-1141647" [ 1022.862215] env[62692]: _type = "Task" [ 1022.862215] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.871832] env[62692]: DEBUG oslo_vmware.api [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141647, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.880331] env[62692]: WARNING nova.network.neutron [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] b34af32c-59b7-4de4-85ef-08bd67c1e2a2 already exists in list: networks containing: ['b34af32c-59b7-4de4-85ef-08bd67c1e2a2']. ignoring it [ 1022.880331] env[62692]: WARNING nova.network.neutron [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] b34af32c-59b7-4de4-85ef-08bd67c1e2a2 already exists in list: networks containing: ['b34af32c-59b7-4de4-85ef-08bd67c1e2a2']. ignoring it [ 1022.890147] env[62692]: DEBUG nova.compute.manager [req-6ad1af7f-1c18-4896-a384-390049580766 req-8bfcb854-5868-4ba0-9447-7791a58ce444 service nova] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Received event network-vif-deleted-8842838a-68c4-4c35-b3dc-ba680c8f3732 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1022.890654] env[62692]: DEBUG nova.compute.manager [req-6ad1af7f-1c18-4896-a384-390049580766 req-8bfcb854-5868-4ba0-9447-7791a58ce444 service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Received event network-vif-plugged-dac9e9e1-d8ad-4ee7-8836-a84a7cb3cf27 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1022.891024] env[62692]: DEBUG oslo_concurrency.lockutils [req-6ad1af7f-1c18-4896-a384-390049580766 req-8bfcb854-5868-4ba0-9447-7791a58ce444 service nova] Acquiring lock "668fb0ea-7eed-4198-943c-8f916ec11368-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.891356] env[62692]: DEBUG oslo_concurrency.lockutils [req-6ad1af7f-1c18-4896-a384-390049580766 req-8bfcb854-5868-4ba0-9447-7791a58ce444 service nova] Lock "668fb0ea-7eed-4198-943c-8f916ec11368-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.892143] env[62692]: DEBUG oslo_concurrency.lockutils [req-6ad1af7f-1c18-4896-a384-390049580766 req-8bfcb854-5868-4ba0-9447-7791a58ce444 service nova] Lock "668fb0ea-7eed-4198-943c-8f916ec11368-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.892143] env[62692]: DEBUG nova.compute.manager [req-6ad1af7f-1c18-4896-a384-390049580766 req-8bfcb854-5868-4ba0-9447-7791a58ce444 service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] No waiting events found dispatching network-vif-plugged-dac9e9e1-d8ad-4ee7-8836-a84a7cb3cf27 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1022.892143] env[62692]: WARNING nova.compute.manager [req-6ad1af7f-1c18-4896-a384-390049580766 req-8bfcb854-5868-4ba0-9447-7791a58ce444 service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Received unexpected event network-vif-plugged-dac9e9e1-d8ad-4ee7-8836-a84a7cb3cf27 for instance with vm_state active and task_state None. [ 1022.893881] env[62692]: DEBUG nova.compute.manager [req-6ad1af7f-1c18-4896-a384-390049580766 req-8bfcb854-5868-4ba0-9447-7791a58ce444 service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Received event network-changed-dac9e9e1-d8ad-4ee7-8836-a84a7cb3cf27 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1022.893881] env[62692]: DEBUG nova.compute.manager [req-6ad1af7f-1c18-4896-a384-390049580766 req-8bfcb854-5868-4ba0-9447-7791a58ce444 service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Refreshing instance network info cache due to event network-changed-dac9e9e1-d8ad-4ee7-8836-a84a7cb3cf27. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1022.893881] env[62692]: DEBUG oslo_concurrency.lockutils [req-6ad1af7f-1c18-4896-a384-390049580766 req-8bfcb854-5868-4ba0-9447-7791a58ce444 service nova] Acquiring lock "refresh_cache-668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.933461] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.940667] env[62692]: DEBUG oslo_vmware.api [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141645, 'name': ReconfigVM_Task, 'duration_secs': 0.174012} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.943902] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Reconfigured VM instance instance-00000021 to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1022.943902] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-538cca8f-edbd-420b-9c54-94c25021861c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.949634] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0bf3700-b303-492f-8151-358e71abc943 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.970658] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33a08315-b6cc-4be5-8760-2a42b8e0130d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.982631] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Reconfiguring VM instance instance-00000021 to attach disk [datastore2] d99737d8-2eb0-40ee-b61e-6c736c84ea59/d99737d8-2eb0-40ee-b61e-6c736c84ea59.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1022.983311] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-98408005-ac88-41be-8a1f-3895165cc321 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.033845] env[62692]: DEBUG oslo_vmware.api [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1023.033845] env[62692]: value = "task-1141648" [ 1023.033845] env[62692]: _type = "Task" [ 1023.033845] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.035642] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06200555-e910-474b-b799-73651f583fc3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.046064] env[62692]: DEBUG oslo_vmware.api [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141646, 'name': PowerOnVM_Task, 'duration_secs': 0.464924} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.049731] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1023.050150] env[62692]: INFO nova.compute.manager [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Took 7.66 seconds to spawn the instance on the hypervisor. [ 1023.050594] env[62692]: DEBUG nova.compute.manager [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1023.057655] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f6e8015-5c68-42a4-bdca-fda93b855ce8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.063171] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-965df19d-8e75-44cb-85ac-238727589bbc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.068550] env[62692]: DEBUG oslo_vmware.api [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141648, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.084039] env[62692]: DEBUG nova.compute.provider_tree [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1023.164581] env[62692]: DEBUG nova.compute.manager [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1023.207297] env[62692]: DEBUG nova.network.neutron [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Successfully created port: 877ab637-ea08-499f-9d34-88ca15e1ed1a {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1023.310787] env[62692]: DEBUG oslo_concurrency.lockutils [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.311530] env[62692]: DEBUG oslo_concurrency.lockutils [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.371455] env[62692]: DEBUG oslo_vmware.api [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141647, 'name': PowerOffVM_Task, 'duration_secs': 0.252843} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.371731] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1023.371904] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1023.372163] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-daa9413c-8f07-44f2-afd1-a956ae09fc27 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.430467] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1023.430858] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1023.430917] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Deleting the datastore file [datastore1] 3f6c9744-a6e3-43f3-8b6b-624ba178747d {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1023.431202] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-be31035c-2ffe-4935-9bc2-7f36a676b423 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.437927] env[62692]: DEBUG oslo_vmware.api [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1023.437927] env[62692]: value = "task-1141650" [ 1023.437927] env[62692]: _type = "Task" [ 1023.437927] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.446051] env[62692]: DEBUG oslo_vmware.api [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141650, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.546208] env[62692]: DEBUG oslo_vmware.api [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141648, 'name': ReconfigVM_Task, 'duration_secs': 0.285737} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.546504] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Reconfigured VM instance instance-00000021 to attach disk [datastore2] d99737d8-2eb0-40ee-b61e-6c736c84ea59/d99737d8-2eb0-40ee-b61e-6c736c84ea59.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1023.546770] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Updating instance 'd99737d8-2eb0-40ee-b61e-6c736c84ea59' progress to 50 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1023.594198] env[62692]: DEBUG nova.scheduler.client.report [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1023.597134] env[62692]: INFO nova.compute.manager [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Took 37.14 seconds to build instance. [ 1023.646438] env[62692]: DEBUG nova.network.neutron [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Updating instance_info_cache with network_info: [{"id": "2deb4259-2f64-4f7c-9289-eba76375335a", "address": "fa:16:3e:bd:3f:b3", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.135", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2deb4259-2f", "ovs_interfaceid": "2deb4259-2f64-4f7c-9289-eba76375335a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e1a45cde-8d94-464d-8c3e-e3f94b62f9e6", "address": "fa:16:3e:aa:dc:ce", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1a45cde-8d", "ovs_interfaceid": "e1a45cde-8d94-464d-8c3e-e3f94b62f9e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "dac9e9e1-d8ad-4ee7-8836-a84a7cb3cf27", "address": "fa:16:3e:69:e0:73", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdac9e9e1-d8", "ovs_interfaceid": "dac9e9e1-d8ad-4ee7-8836-a84a7cb3cf27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.812872] env[62692]: DEBUG nova.compute.manager [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1023.948528] env[62692]: DEBUG oslo_vmware.api [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141650, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147698} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.948804] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1023.948996] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1023.949204] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1023.949379] env[62692]: INFO nova.compute.manager [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1023.949697] env[62692]: DEBUG oslo.service.loopingcall [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1023.949895] env[62692]: DEBUG nova.compute.manager [-] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1023.949989] env[62692]: DEBUG nova.network.neutron [-] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1024.054059] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-236c2b61-d55a-4ef8-8416-8a57fc4d8b6c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.075254] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d19fa5f-39f0-4d30-8016-c9ad2f1fe29b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.093740] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Updating instance 'd99737d8-2eb0-40ee-b61e-6c736c84ea59' progress to 67 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1024.099546] env[62692]: DEBUG oslo_concurrency.lockutils [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.943s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.101740] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0b9c4754-a32a-4492-8507-dda0db5d598c tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Lock "1b753882-9118-479c-aa3e-cbb8bdf2e086" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.661s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.102223] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.404s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.102449] env[62692]: DEBUG nova.objects.instance [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lazy-loading 'resources' on Instance uuid f3e8be1c-d952-49b7-86a8-d0a6c56ee096 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1024.127376] env[62692]: INFO nova.scheduler.client.report [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Deleted allocations for instance b23d3484-4fc5-46b2-8e8c-bdaab087f046 [ 1024.149898] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "refresh_cache-668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.150231] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.150394] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.150701] env[62692]: DEBUG oslo_concurrency.lockutils [req-6ad1af7f-1c18-4896-a384-390049580766 req-8bfcb854-5868-4ba0-9447-7791a58ce444 service nova] Acquired lock "refresh_cache-668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.150891] env[62692]: DEBUG nova.network.neutron [req-6ad1af7f-1c18-4896-a384-390049580766 req-8bfcb854-5868-4ba0-9447-7791a58ce444 service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Refreshing network info cache for port dac9e9e1-d8ad-4ee7-8836-a84a7cb3cf27 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1024.154545] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-100c6d7a-f529-496a-b803-0d0a27a48d25 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.172850] env[62692]: DEBUG nova.virt.hardware [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1024.173029] env[62692]: DEBUG nova.virt.hardware [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1024.173152] env[62692]: DEBUG nova.virt.hardware [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1024.173346] env[62692]: DEBUG nova.virt.hardware [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1024.173500] env[62692]: DEBUG nova.virt.hardware [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1024.173659] env[62692]: DEBUG nova.virt.hardware [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1024.174027] env[62692]: DEBUG nova.virt.hardware [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1024.174152] env[62692]: DEBUG nova.virt.hardware [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1024.174209] env[62692]: DEBUG nova.virt.hardware [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1024.174873] env[62692]: DEBUG nova.virt.hardware [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1024.174873] env[62692]: DEBUG nova.virt.hardware [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1024.181060] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Reconfiguring VM to attach interface {{(pid=62692) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1024.182340] env[62692]: DEBUG nova.compute.manager [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1024.185520] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b23c6677-2247-4915-94ef-bf7bd85a0baa {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.208167] env[62692]: DEBUG oslo_vmware.api [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1024.208167] env[62692]: value = "task-1141651" [ 1024.208167] env[62692]: _type = "Task" [ 1024.208167] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.218762] env[62692]: DEBUG oslo_vmware.api [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141651, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.224861] env[62692]: DEBUG nova.virt.hardware [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1024.225118] env[62692]: DEBUG nova.virt.hardware [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1024.225288] env[62692]: DEBUG nova.virt.hardware [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1024.225481] env[62692]: DEBUG nova.virt.hardware [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1024.225634] env[62692]: DEBUG nova.virt.hardware [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1024.225784] env[62692]: DEBUG nova.virt.hardware [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1024.225995] env[62692]: DEBUG nova.virt.hardware [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1024.226176] env[62692]: DEBUG nova.virt.hardware [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1024.226348] env[62692]: DEBUG nova.virt.hardware [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1024.226511] env[62692]: DEBUG nova.virt.hardware [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1024.226690] env[62692]: DEBUG nova.virt.hardware [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1024.227710] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c732ff4f-cd0c-4187-bbb6-32e67fb841d6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.235083] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d7a5e5c-b2fb-4d07-a7b7-3b16ffa4d4cd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.333397] env[62692]: DEBUG oslo_concurrency.lockutils [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.636172] env[62692]: DEBUG oslo_concurrency.lockutils [None req-568282c0-fec8-4cdb-a58e-bbce5b7f0c4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "b23d3484-4fc5-46b2-8e8c-bdaab087f046" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.450s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.640661] env[62692]: DEBUG nova.network.neutron [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Port 7f2cabdf-cbda-4cf7-a518-13f619716d12 binding to destination host cpu-1 is already ACTIVE {{(pid=62692) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1024.700284] env[62692]: DEBUG nova.network.neutron [-] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.727886] env[62692]: DEBUG oslo_vmware.api [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141651, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.899504] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0206827b-162d-4910-99ae-5a1aaaf14043 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.907471] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38d93ebe-686c-452b-a6d1-4b5a60a8a621 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.939609] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f1aec23-8713-474b-b19e-ad380b938497 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.947527] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-324e9702-2055-47a3-940f-ec188e0ed0a0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.961242] env[62692]: DEBUG nova.compute.provider_tree [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1025.024519] env[62692]: DEBUG nova.network.neutron [req-6ad1af7f-1c18-4896-a384-390049580766 req-8bfcb854-5868-4ba0-9447-7791a58ce444 service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Updated VIF entry in instance network info cache for port dac9e9e1-d8ad-4ee7-8836-a84a7cb3cf27. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1025.025150] env[62692]: DEBUG nova.network.neutron [req-6ad1af7f-1c18-4896-a384-390049580766 req-8bfcb854-5868-4ba0-9447-7791a58ce444 service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Updating instance_info_cache with network_info: [{"id": "2deb4259-2f64-4f7c-9289-eba76375335a", "address": "fa:16:3e:bd:3f:b3", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.135", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2deb4259-2f", "ovs_interfaceid": "2deb4259-2f64-4f7c-9289-eba76375335a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e1a45cde-8d94-464d-8c3e-e3f94b62f9e6", "address": "fa:16:3e:aa:dc:ce", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1a45cde-8d", "ovs_interfaceid": "e1a45cde-8d94-464d-8c3e-e3f94b62f9e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "dac9e9e1-d8ad-4ee7-8836-a84a7cb3cf27", "address": "fa:16:3e:69:e0:73", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdac9e9e1-d8", "ovs_interfaceid": "dac9e9e1-d8ad-4ee7-8836-a84a7cb3cf27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.075018] env[62692]: DEBUG nova.network.neutron [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Successfully updated port: 877ab637-ea08-499f-9d34-88ca15e1ed1a {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1025.127511] env[62692]: DEBUG nova.compute.manager [req-6a61ba0c-e6ea-4680-85d8-80b8190ab1f7 req-acc66664-2fcd-4549-a9da-a82e63574c23 service nova] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Received event network-vif-deleted-2cb4b52c-0d89-4402-ae48-dc6f226c557c {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1025.202904] env[62692]: INFO nova.compute.manager [-] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Took 1.25 seconds to deallocate network for instance. [ 1025.220431] env[62692]: DEBUG oslo_vmware.api [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141651, 'name': ReconfigVM_Task, 'duration_secs': 0.608449} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.220994] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.221236] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Reconfigured VM to attach interface {{(pid=62692) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1025.362947] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquiring lock "af260f9a-4e9b-4af1-90da-c619c4755eca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.363219] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Lock "af260f9a-4e9b-4af1-90da-c619c4755eca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.463916] env[62692]: DEBUG nova.scheduler.client.report [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1025.527515] env[62692]: DEBUG oslo_concurrency.lockutils [req-6ad1af7f-1c18-4896-a384-390049580766 req-8bfcb854-5868-4ba0-9447-7791a58ce444 service nova] Releasing lock "refresh_cache-668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.579115] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "refresh_cache-e8f645da-3b4a-4f21-9a17-aebdc8d112df" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.579270] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquired lock "refresh_cache-e8f645da-3b4a-4f21-9a17-aebdc8d112df" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.579436] env[62692]: DEBUG nova.network.neutron [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1025.664553] env[62692]: DEBUG oslo_concurrency.lockutils [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "d99737d8-2eb0-40ee-b61e-6c736c84ea59-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.664959] env[62692]: DEBUG oslo_concurrency.lockutils [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "d99737d8-2eb0-40ee-b61e-6c736c84ea59-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.665025] env[62692]: DEBUG oslo_concurrency.lockutils [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "d99737d8-2eb0-40ee-b61e-6c736c84ea59-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.712580] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.729221] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d1d89938-502c-430b-8bb7-5160202e88a2 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "interface-668fb0ea-7eed-4198-943c-8f916ec11368-dac9e9e1-d8ad-4ee7-8836-a84a7cb3cf27" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.534s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.866186] env[62692]: DEBUG nova.compute.manager [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1025.969326] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.867s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.971578] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.161s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.973358] env[62692]: DEBUG nova.objects.instance [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lazy-loading 'resources' on Instance uuid 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1025.991549] env[62692]: INFO nova.scheduler.client.report [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Deleted allocations for instance f3e8be1c-d952-49b7-86a8-d0a6c56ee096 [ 1026.092400] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "a1951dba-ec5a-4849-9e22-37fba08ec60c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.092642] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "a1951dba-ec5a-4849-9e22-37fba08ec60c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.111660] env[62692]: DEBUG nova.network.neutron [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1026.250085] env[62692]: DEBUG nova.network.neutron [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Updating instance_info_cache with network_info: [{"id": "877ab637-ea08-499f-9d34-88ca15e1ed1a", "address": "fa:16:3e:73:e7:26", "network": {"id": "2febb660-329d-403c-8c2c-f11c0f5f547c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1806650357-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37921f67664544478cb2568def068b11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap877ab637-ea", "ovs_interfaceid": "877ab637-ea08-499f-9d34-88ca15e1ed1a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.385439] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.499772] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0960b5c4-b09b-4cfd-8dfb-fb988b4a926c tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "f3e8be1c-d952-49b7-86a8-d0a6c56ee096" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.994s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.595529] env[62692]: DEBUG nova.compute.manager [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1026.698425] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e948d043-c064-40ea-907b-9d1acdcaf9b9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.703192] env[62692]: DEBUG oslo_concurrency.lockutils [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "refresh_cache-d99737d8-2eb0-40ee-b61e-6c736c84ea59" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.703410] env[62692]: DEBUG oslo_concurrency.lockutils [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired lock "refresh_cache-d99737d8-2eb0-40ee-b61e-6c736c84ea59" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.703561] env[62692]: DEBUG nova.network.neutron [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1026.708261] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c76159a-3a32-47de-a487-b001a94152d7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.745666] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a83508ea-d80f-4dd4-a4ef-326fa8583546 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.754239] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Releasing lock "refresh_cache-e8f645da-3b4a-4f21-9a17-aebdc8d112df" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.754382] env[62692]: DEBUG nova.compute.manager [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Instance network_info: |[{"id": "877ab637-ea08-499f-9d34-88ca15e1ed1a", "address": "fa:16:3e:73:e7:26", "network": {"id": "2febb660-329d-403c-8c2c-f11c0f5f547c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1806650357-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37921f67664544478cb2568def068b11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap877ab637-ea", "ovs_interfaceid": "877ab637-ea08-499f-9d34-88ca15e1ed1a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1026.754850] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:73:e7:26', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '877ab637-ea08-499f-9d34-88ca15e1ed1a', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1026.764012] env[62692]: DEBUG oslo.service.loopingcall [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1026.765148] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-007990ca-e366-4ff2-a155-923a61d30072 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.769539] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1026.769785] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-077d06ca-f304-40b8-b6d7-f796bd2f0dfe {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.794191] env[62692]: DEBUG nova.compute.provider_tree [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1026.797196] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1026.797196] env[62692]: value = "task-1141652" [ 1026.797196] env[62692]: _type = "Task" [ 1026.797196] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.805907] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141652, 'name': CreateVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.124796] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.205149] env[62692]: DEBUG nova.compute.manager [req-1a9f9a49-4522-444c-87ac-237fab1ca0a2 req-e19dfe5a-30fd-4ea5-b10e-903058ca04c2 service nova] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Received event network-vif-plugged-877ab637-ea08-499f-9d34-88ca15e1ed1a {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1027.205385] env[62692]: DEBUG oslo_concurrency.lockutils [req-1a9f9a49-4522-444c-87ac-237fab1ca0a2 req-e19dfe5a-30fd-4ea5-b10e-903058ca04c2 service nova] Acquiring lock "e8f645da-3b4a-4f21-9a17-aebdc8d112df-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.205602] env[62692]: DEBUG oslo_concurrency.lockutils [req-1a9f9a49-4522-444c-87ac-237fab1ca0a2 req-e19dfe5a-30fd-4ea5-b10e-903058ca04c2 service nova] Lock "e8f645da-3b4a-4f21-9a17-aebdc8d112df-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.205783] env[62692]: DEBUG oslo_concurrency.lockutils [req-1a9f9a49-4522-444c-87ac-237fab1ca0a2 req-e19dfe5a-30fd-4ea5-b10e-903058ca04c2 service nova] Lock "e8f645da-3b4a-4f21-9a17-aebdc8d112df-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.205998] env[62692]: DEBUG nova.compute.manager [req-1a9f9a49-4522-444c-87ac-237fab1ca0a2 req-e19dfe5a-30fd-4ea5-b10e-903058ca04c2 service nova] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] No waiting events found dispatching network-vif-plugged-877ab637-ea08-499f-9d34-88ca15e1ed1a {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1027.206158] env[62692]: WARNING nova.compute.manager [req-1a9f9a49-4522-444c-87ac-237fab1ca0a2 req-e19dfe5a-30fd-4ea5-b10e-903058ca04c2 service nova] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Received unexpected event network-vif-plugged-877ab637-ea08-499f-9d34-88ca15e1ed1a for instance with vm_state building and task_state spawning. [ 1027.206330] env[62692]: DEBUG nova.compute.manager [req-1a9f9a49-4522-444c-87ac-237fab1ca0a2 req-e19dfe5a-30fd-4ea5-b10e-903058ca04c2 service nova] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Received event network-changed-877ab637-ea08-499f-9d34-88ca15e1ed1a {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1027.206489] env[62692]: DEBUG nova.compute.manager [req-1a9f9a49-4522-444c-87ac-237fab1ca0a2 req-e19dfe5a-30fd-4ea5-b10e-903058ca04c2 service nova] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Refreshing instance network info cache due to event network-changed-877ab637-ea08-499f-9d34-88ca15e1ed1a. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1027.206677] env[62692]: DEBUG oslo_concurrency.lockutils [req-1a9f9a49-4522-444c-87ac-237fab1ca0a2 req-e19dfe5a-30fd-4ea5-b10e-903058ca04c2 service nova] Acquiring lock "refresh_cache-e8f645da-3b4a-4f21-9a17-aebdc8d112df" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.206817] env[62692]: DEBUG oslo_concurrency.lockutils [req-1a9f9a49-4522-444c-87ac-237fab1ca0a2 req-e19dfe5a-30fd-4ea5-b10e-903058ca04c2 service nova] Acquired lock "refresh_cache-e8f645da-3b4a-4f21-9a17-aebdc8d112df" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.206991] env[62692]: DEBUG nova.network.neutron [req-1a9f9a49-4522-444c-87ac-237fab1ca0a2 req-e19dfe5a-30fd-4ea5-b10e-903058ca04c2 service nova] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Refreshing network info cache for port 877ab637-ea08-499f-9d34-88ca15e1ed1a {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1027.298291] env[62692]: DEBUG nova.scheduler.client.report [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1027.312375] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141652, 'name': CreateVM_Task, 'duration_secs': 0.390856} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.312566] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1027.314106] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.314106] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.314106] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1027.314319] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62ed182c-9421-4201-8649-cd6fa3b13ab6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.319803] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1027.319803] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]525be74c-1d18-63e8-0ed7-0182e573a9be" [ 1027.319803] env[62692]: _type = "Task" [ 1027.319803] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.329099] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]525be74c-1d18-63e8-0ed7-0182e573a9be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.508259] env[62692]: DEBUG nova.network.neutron [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Updating instance_info_cache with network_info: [{"id": "7f2cabdf-cbda-4cf7-a518-13f619716d12", "address": "fa:16:3e:6b:22:0c", "network": {"id": "f70727d6-a9aa-44b0-a244-2533b3f96571", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-284105837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1037e5d436bd429391159b87a712f6e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f2cabdf-cb", "ovs_interfaceid": "7f2cabdf-cbda-4cf7-a518-13f619716d12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.735832] env[62692]: DEBUG oslo_concurrency.lockutils [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "interface-668fb0ea-7eed-4198-943c-8f916ec11368-e1a45cde-8d94-464d-8c3e-e3f94b62f9e6" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.736282] env[62692]: DEBUG oslo_concurrency.lockutils [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "interface-668fb0ea-7eed-4198-943c-8f916ec11368-e1a45cde-8d94-464d-8c3e-e3f94b62f9e6" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.809245] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.838s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.811325] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.646s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.811541] env[62692]: DEBUG nova.objects.instance [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lazy-loading 'resources' on Instance uuid f6085887-a5d3-4096-be38-d843bb54304a {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.830728] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]525be74c-1d18-63e8-0ed7-0182e573a9be, 'name': SearchDatastore_Task, 'duration_secs': 0.010132} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.833091] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.833348] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1027.833589] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.833741] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.833921] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1027.834865] env[62692]: INFO nova.scheduler.client.report [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Deleted allocations for instance 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19 [ 1027.835774] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-564be6e4-7cc4-43b1-95e6-6195a52784d4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.846913] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1027.847120] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1027.847886] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59143830-bbe5-4bf3-9928-ba0148228b84 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.853168] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1027.853168] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52f3a9ce-1969-9e8b-b3c4-1f0f5d35dc08" [ 1027.853168] env[62692]: _type = "Task" [ 1027.853168] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.862983] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52f3a9ce-1969-9e8b-b3c4-1f0f5d35dc08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.953432] env[62692]: DEBUG nova.network.neutron [req-1a9f9a49-4522-444c-87ac-237fab1ca0a2 req-e19dfe5a-30fd-4ea5-b10e-903058ca04c2 service nova] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Updated VIF entry in instance network info cache for port 877ab637-ea08-499f-9d34-88ca15e1ed1a. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1027.953432] env[62692]: DEBUG nova.network.neutron [req-1a9f9a49-4522-444c-87ac-237fab1ca0a2 req-e19dfe5a-30fd-4ea5-b10e-903058ca04c2 service nova] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Updating instance_info_cache with network_info: [{"id": "877ab637-ea08-499f-9d34-88ca15e1ed1a", "address": "fa:16:3e:73:e7:26", "network": {"id": "2febb660-329d-403c-8c2c-f11c0f5f547c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1806650357-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37921f67664544478cb2568def068b11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap877ab637-ea", "ovs_interfaceid": "877ab637-ea08-499f-9d34-88ca15e1ed1a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.010646] env[62692]: DEBUG oslo_concurrency.lockutils [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Releasing lock "refresh_cache-d99737d8-2eb0-40ee-b61e-6c736c84ea59" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.189049] env[62692]: DEBUG oslo_concurrency.lockutils [None req-41b46abf-f915-48a1-b336-08f3b7a3677b tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "754b3c65-1e4b-49d2-8980-095d975edb01" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.189345] env[62692]: DEBUG oslo_concurrency.lockutils [None req-41b46abf-f915-48a1-b336-08f3b7a3677b tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "754b3c65-1e4b-49d2-8980-095d975edb01" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.239144] env[62692]: DEBUG oslo_concurrency.lockutils [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.239362] env[62692]: DEBUG oslo_concurrency.lockutils [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.240293] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524dfaaf-beef-4709-83af-d54f3f25cc5c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.258682] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0652872-3d62-4ff3-a829-be4428abffd5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.288806] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Reconfiguring VM to detach interface {{(pid=62692) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1028.289077] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a03972c1-cd6e-4db8-ba47-8f13f2ec6802 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.306958] env[62692]: DEBUG oslo_vmware.api [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1028.306958] env[62692]: value = "task-1141653" [ 1028.306958] env[62692]: _type = "Task" [ 1028.306958] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.316589] env[62692]: DEBUG oslo_vmware.api [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141653, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.346267] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96da420b-3a0e-45dd-b73a-cd13363e3452 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.665s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.364213] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52f3a9ce-1969-9e8b-b3c4-1f0f5d35dc08, 'name': SearchDatastore_Task, 'duration_secs': 0.009101} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.365609] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4e62cc5-0a0c-4b56-a3fc-2d5314bf94aa {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.371439] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1028.371439] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52de4508-2454-0faa-5ead-35d9196dcaa4" [ 1028.371439] env[62692]: _type = "Task" [ 1028.371439] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.381671] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52de4508-2454-0faa-5ead-35d9196dcaa4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.456862] env[62692]: DEBUG oslo_concurrency.lockutils [req-1a9f9a49-4522-444c-87ac-237fab1ca0a2 req-e19dfe5a-30fd-4ea5-b10e-903058ca04c2 service nova] Releasing lock "refresh_cache-e8f645da-3b4a-4f21-9a17-aebdc8d112df" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.530962] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f34a9696-1bb1-4485-96d7-f05eeef8b565 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.551455] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d0cade6-6041-406f-a89f-cb3df56791aa {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.558964] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Updating instance 'd99737d8-2eb0-40ee-b61e-6c736c84ea59' progress to 83 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1028.575416] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbf71ff5-6714-467f-a6a6-e13d489c0cd2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.582710] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82ced3c3-1c3d-4abb-ac5f-496c59f0abb2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.614785] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5e8eb9b-2ba6-4e97-b11d-309ea7b91f7f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.622698] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63965553-ec36-4cdd-b128-c93188e1dd26 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.636601] env[62692]: DEBUG nova.compute.provider_tree [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.692560] env[62692]: INFO nova.compute.manager [None req-41b46abf-f915-48a1-b336-08f3b7a3677b tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Detaching volume 57d15e0c-bb59-47d1-abf1-252dc8cb294a [ 1028.725518] env[62692]: INFO nova.virt.block_device [None req-41b46abf-f915-48a1-b336-08f3b7a3677b tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Attempting to driver detach volume 57d15e0c-bb59-47d1-abf1-252dc8cb294a from mountpoint /dev/sdb [ 1028.726096] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-41b46abf-f915-48a1-b336-08f3b7a3677b tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Volume detach. Driver type: vmdk {{(pid=62692) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1028.726296] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-41b46abf-f915-48a1-b336-08f3b7a3677b tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-248974', 'volume_id': '57d15e0c-bb59-47d1-abf1-252dc8cb294a', 'name': 'volume-57d15e0c-bb59-47d1-abf1-252dc8cb294a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '754b3c65-1e4b-49d2-8980-095d975edb01', 'attached_at': '', 'detached_at': '', 'volume_id': '57d15e0c-bb59-47d1-abf1-252dc8cb294a', 'serial': '57d15e0c-bb59-47d1-abf1-252dc8cb294a'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1028.727328] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f5a3104-87f4-459d-ac74-2d52b3f3264a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.752155] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e39fae8d-07bf-49ce-a1b5-6ff9b9613f11 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.759969] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3b3e4f5-bb43-40dc-8278-a757409de066 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.780185] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52096722-96c0-48d3-969a-8964e65d4bb8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.795402] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-41b46abf-f915-48a1-b336-08f3b7a3677b tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] The volume has not been displaced from its original location: [datastore1] volume-57d15e0c-bb59-47d1-abf1-252dc8cb294a/volume-57d15e0c-bb59-47d1-abf1-252dc8cb294a.vmdk. No consolidation needed. {{(pid=62692) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1028.800601] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-41b46abf-f915-48a1-b336-08f3b7a3677b tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Reconfiguring VM instance instance-0000003b to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1028.800888] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d04f7f1-284c-4a24-83ac-dad42a0b9089 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.823159] env[62692]: DEBUG oslo_vmware.api [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141653, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.824349] env[62692]: DEBUG oslo_vmware.api [None req-41b46abf-f915-48a1-b336-08f3b7a3677b tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1028.824349] env[62692]: value = "task-1141655" [ 1028.824349] env[62692]: _type = "Task" [ 1028.824349] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.831640] env[62692]: DEBUG oslo_vmware.api [None req-41b46abf-f915-48a1-b336-08f3b7a3677b tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141655, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.883871] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52de4508-2454-0faa-5ead-35d9196dcaa4, 'name': SearchDatastore_Task, 'duration_secs': 0.009344} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.884466] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.884509] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] e8f645da-3b4a-4f21-9a17-aebdc8d112df/e8f645da-3b4a-4f21-9a17-aebdc8d112df.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1028.884786] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ca9ad6c6-11a7-477c-85dc-840f99a4cc32 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.890994] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1028.890994] env[62692]: value = "task-1141656" [ 1028.890994] env[62692]: _type = "Task" [ 1028.890994] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.899327] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141656, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.069285] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1029.069795] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-991419e2-f811-488f-8033-1512a52651b2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.080197] env[62692]: DEBUG oslo_vmware.api [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1029.080197] env[62692]: value = "task-1141657" [ 1029.080197] env[62692]: _type = "Task" [ 1029.080197] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.093118] env[62692]: DEBUG oslo_vmware.api [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141657, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.139896] env[62692]: DEBUG nova.scheduler.client.report [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1029.324254] env[62692]: DEBUG oslo_vmware.api [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141653, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.335499] env[62692]: DEBUG oslo_vmware.api [None req-41b46abf-f915-48a1-b336-08f3b7a3677b tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141655, 'name': ReconfigVM_Task, 'duration_secs': 0.237984} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.335796] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-41b46abf-f915-48a1-b336-08f3b7a3677b tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Reconfigured VM instance instance-0000003b to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1029.340704] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e0e53843-c923-4168-85b2-2781182704a9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.358419] env[62692]: DEBUG oslo_vmware.api [None req-41b46abf-f915-48a1-b336-08f3b7a3677b tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1029.358419] env[62692]: value = "task-1141658" [ 1029.358419] env[62692]: _type = "Task" [ 1029.358419] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.367401] env[62692]: DEBUG oslo_vmware.api [None req-41b46abf-f915-48a1-b336-08f3b7a3677b tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141658, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.402383] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141656, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.590774] env[62692]: DEBUG oslo_vmware.api [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141657, 'name': PowerOnVM_Task, 'duration_secs': 0.428257} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.591318] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1029.591318] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-34c2b590-6632-4de0-8888-eff46afe4680 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Updating instance 'd99737d8-2eb0-40ee-b61e-6c736c84ea59' progress to 100 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1029.648200] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.837s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.650664] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.340s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.650913] env[62692]: DEBUG nova.objects.instance [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lazy-loading 'resources' on Instance uuid 9b8e830c-61b7-4dd2-8324-d3a96eec1465 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.675130] env[62692]: INFO nova.scheduler.client.report [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Deleted allocations for instance f6085887-a5d3-4096-be38-d843bb54304a [ 1029.823878] env[62692]: DEBUG oslo_vmware.api [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141653, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.868094] env[62692]: DEBUG oslo_vmware.api [None req-41b46abf-f915-48a1-b336-08f3b7a3677b tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141658, 'name': ReconfigVM_Task, 'duration_secs': 0.160827} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.868387] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-41b46abf-f915-48a1-b336-08f3b7a3677b tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-248974', 'volume_id': '57d15e0c-bb59-47d1-abf1-252dc8cb294a', 'name': 'volume-57d15e0c-bb59-47d1-abf1-252dc8cb294a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '754b3c65-1e4b-49d2-8980-095d975edb01', 'attached_at': '', 'detached_at': '', 'volume_id': '57d15e0c-bb59-47d1-abf1-252dc8cb294a', 'serial': '57d15e0c-bb59-47d1-abf1-252dc8cb294a'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1029.901866] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141656, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.53918} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.902150] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] e8f645da-3b4a-4f21-9a17-aebdc8d112df/e8f645da-3b4a-4f21-9a17-aebdc8d112df.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1029.902376] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1029.902632] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4685f917-5aba-4cb7-8e64-a3b780691fcd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.909531] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1029.909531] env[62692]: value = "task-1141659" [ 1029.909531] env[62692]: _type = "Task" [ 1029.909531] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.917878] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141659, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.185789] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d7add4c2-34e1-4b1a-ab1a-39f2a46b9cd8 tempest-ImagesTestJSON-1190674654 tempest-ImagesTestJSON-1190674654-project-member] Lock "f6085887-a5d3-4096-be38-d843bb54304a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.932s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.327018] env[62692]: DEBUG oslo_vmware.api [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141653, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.391786] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5920bc94-804b-4ac4-baf1-b4b6fe11e723 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.399591] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40494601-efe7-46f8-a72b-04d7cbb9725f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.431218] env[62692]: DEBUG nova.objects.instance [None req-41b46abf-f915-48a1-b336-08f3b7a3677b tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lazy-loading 'flavor' on Instance uuid 754b3c65-1e4b-49d2-8980-095d975edb01 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1030.436585] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e52b98bb-da23-4418-b32a-7452e046f11c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.445065] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141659, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.449715] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b9942ce-4b65-49b9-a56a-2e67a08a4676 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.465467] env[62692]: DEBUG nova.compute.provider_tree [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1030.836874] env[62692]: DEBUG oslo_vmware.api [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141653, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.952789] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141659, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.719775} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.955128] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1030.956616] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-410ad4b7-d753-4908-b92d-f810cb14746e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.973040] env[62692]: DEBUG nova.scheduler.client.report [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1030.987577] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] e8f645da-3b4a-4f21-9a17-aebdc8d112df/e8f645da-3b4a-4f21-9a17-aebdc8d112df.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1030.990214] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-72448296-cd9f-4d37-84af-9a64a4956c93 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.014023] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1031.014023] env[62692]: value = "task-1141660" [ 1031.014023] env[62692]: _type = "Task" [ 1031.014023] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.021425] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141660, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.330514] env[62692]: DEBUG oslo_vmware.api [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141653, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.453274] env[62692]: DEBUG oslo_concurrency.lockutils [None req-25e44e32-4bcf-45a3-9769-eca012e35bf1 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "d99737d8-2eb0-40ee-b61e-6c736c84ea59" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.453551] env[62692]: DEBUG oslo_concurrency.lockutils [None req-25e44e32-4bcf-45a3-9769-eca012e35bf1 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "d99737d8-2eb0-40ee-b61e-6c736c84ea59" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.453746] env[62692]: DEBUG nova.compute.manager [None req-25e44e32-4bcf-45a3-9769-eca012e35bf1 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Going to confirm migration 3 {{(pid=62692) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1031.455638] env[62692]: DEBUG oslo_concurrency.lockutils [None req-41b46abf-f915-48a1-b336-08f3b7a3677b tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "754b3c65-1e4b-49d2-8980-095d975edb01" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.266s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.491039] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.840s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.493724] env[62692]: DEBUG oslo_concurrency.lockutils [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.226s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.494009] env[62692]: DEBUG nova.objects.instance [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lazy-loading 'resources' on Instance uuid 250b9cc2-706e-4fbf-887c-1c80fc27a0ed {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1031.524947] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141660, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.528266] env[62692]: INFO nova.scheduler.client.report [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Deleted allocations for instance 9b8e830c-61b7-4dd2-8324-d3a96eec1465 [ 1031.831640] env[62692]: DEBUG oslo_vmware.api [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141653, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.997236] env[62692]: DEBUG nova.objects.instance [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lazy-loading 'numa_topology' on Instance uuid 250b9cc2-706e-4fbf-887c-1c80fc27a0ed {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1032.000974] env[62692]: DEBUG oslo_concurrency.lockutils [None req-25e44e32-4bcf-45a3-9769-eca012e35bf1 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "refresh_cache-d99737d8-2eb0-40ee-b61e-6c736c84ea59" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.001435] env[62692]: DEBUG oslo_concurrency.lockutils [None req-25e44e32-4bcf-45a3-9769-eca012e35bf1 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired lock "refresh_cache-d99737d8-2eb0-40ee-b61e-6c736c84ea59" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.001435] env[62692]: DEBUG nova.network.neutron [None req-25e44e32-4bcf-45a3-9769-eca012e35bf1 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1032.001586] env[62692]: DEBUG nova.objects.instance [None req-25e44e32-4bcf-45a3-9769-eca012e35bf1 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lazy-loading 'info_cache' on Instance uuid d99737d8-2eb0-40ee-b61e-6c736c84ea59 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1032.026406] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141660, 'name': ReconfigVM_Task, 'duration_secs': 0.883773} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.026711] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Reconfigured VM instance instance-00000059 to attach disk [datastore1] e8f645da-3b4a-4f21-9a17-aebdc8d112df/e8f645da-3b4a-4f21-9a17-aebdc8d112df.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1032.027372] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e0a2b212-995f-4265-8742-31dfa07c11e2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.036802] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1032.036802] env[62692]: value = "task-1141661" [ 1032.036802] env[62692]: _type = "Task" [ 1032.036802] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.037246] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0ca0c150-d178-4808-ada5-e49919a1fb99 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "9b8e830c-61b7-4dd2-8324-d3a96eec1465" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.649s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.049467] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141661, 'name': Rename_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.332359] env[62692]: DEBUG oslo_vmware.api [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141653, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.501109] env[62692]: DEBUG nova.objects.base [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Object Instance<250b9cc2-706e-4fbf-887c-1c80fc27a0ed> lazy-loaded attributes: resources,numa_topology {{(pid=62692) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1032.535437] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "754b3c65-1e4b-49d2-8980-095d975edb01" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.535701] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "754b3c65-1e4b-49d2-8980-095d975edb01" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.535915] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "754b3c65-1e4b-49d2-8980-095d975edb01-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.536129] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "754b3c65-1e4b-49d2-8980-095d975edb01-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.536309] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "754b3c65-1e4b-49d2-8980-095d975edb01-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.538565] env[62692]: INFO nova.compute.manager [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Terminating instance [ 1032.540433] env[62692]: DEBUG nova.compute.manager [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1032.540707] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1032.544620] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa79a3c3-b3a6-476b-a017-df07ab5f4fc8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.555128] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141661, 'name': Rename_Task, 'duration_secs': 0.367413} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.557153] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1032.557449] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1032.557686] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3ad01fdc-65aa-4548-9b45-e0f5860e929e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.559098] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cb393159-abf1-4456-83f9-e0e524fe2088 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.566931] env[62692]: DEBUG oslo_vmware.api [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1032.566931] env[62692]: value = "task-1141663" [ 1032.566931] env[62692]: _type = "Task" [ 1032.566931] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.568170] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1032.568170] env[62692]: value = "task-1141662" [ 1032.568170] env[62692]: _type = "Task" [ 1032.568170] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.582465] env[62692]: DEBUG oslo_vmware.api [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141663, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.585578] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141662, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.728953] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4dbf690-e390-4fc6-b444-50c7ef02febf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.737150] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e501f26-8f07-4e7f-836b-010a50942fdd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.770319] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca64d47d-1d34-49c8-aadf-773e3ecdd4d3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.780301] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3045750-e312-44be-b7da-da762e99825a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.784661] env[62692]: DEBUG oslo_concurrency.lockutils [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "5a08fa8a-f9fe-4879-bb7b-baa04097df6c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.784897] env[62692]: DEBUG oslo_concurrency.lockutils [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "5a08fa8a-f9fe-4879-bb7b-baa04097df6c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.785109] env[62692]: DEBUG oslo_concurrency.lockutils [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "5a08fa8a-f9fe-4879-bb7b-baa04097df6c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.786989] env[62692]: DEBUG oslo_concurrency.lockutils [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "5a08fa8a-f9fe-4879-bb7b-baa04097df6c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.786989] env[62692]: DEBUG oslo_concurrency.lockutils [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "5a08fa8a-f9fe-4879-bb7b-baa04097df6c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.787997] env[62692]: INFO nova.compute.manager [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Terminating instance [ 1032.797680] env[62692]: DEBUG nova.compute.provider_tree [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1032.799303] env[62692]: DEBUG nova.compute.manager [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1032.799510] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1032.800936] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1adbd8bf-72c5-475b-b401-69132be791a0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.807820] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1032.808123] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f4547eaa-b1d4-41a6-a0d7-1885f00d9f0b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.815131] env[62692]: DEBUG oslo_vmware.api [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 1032.815131] env[62692]: value = "task-1141664" [ 1032.815131] env[62692]: _type = "Task" [ 1032.815131] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.823261] env[62692]: DEBUG oslo_vmware.api [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141664, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.831607] env[62692]: DEBUG oslo_vmware.api [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141653, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.083587] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141662, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.083950] env[62692]: DEBUG oslo_vmware.api [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141663, 'name': PowerOffVM_Task, 'duration_secs': 0.313741} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.084268] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1033.084553] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1033.084830] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-48224f57-0e5b-4e20-a97a-b339164e8723 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.305418] env[62692]: DEBUG nova.scheduler.client.report [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1033.326645] env[62692]: DEBUG oslo_vmware.api [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141664, 'name': PowerOffVM_Task, 'duration_secs': 0.24608} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.330154] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1033.330399] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1033.331526] env[62692]: DEBUG nova.network.neutron [None req-25e44e32-4bcf-45a3-9769-eca012e35bf1 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Updating instance_info_cache with network_info: [{"id": "7f2cabdf-cbda-4cf7-a518-13f619716d12", "address": "fa:16:3e:6b:22:0c", "network": {"id": "f70727d6-a9aa-44b0-a244-2533b3f96571", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-284105837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1037e5d436bd429391159b87a712f6e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f2cabdf-cb", "ovs_interfaceid": "7f2cabdf-cbda-4cf7-a518-13f619716d12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.337157] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4e15e2db-ec9f-48f3-99f3-4af68eee5eb2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.346104] env[62692]: DEBUG oslo_vmware.api [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141653, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.472023] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.472416] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.472681] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.472960] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.473212] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.475481] env[62692]: INFO nova.compute.manager [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Terminating instance [ 1033.477538] env[62692]: DEBUG nova.compute.manager [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1033.477779] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1033.478751] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-751b7fa9-d4be-4065-8f90-c7a78acc7343 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.489546] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1033.489887] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-69ff754a-66ba-4322-a2f6-576de499cc75 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.496879] env[62692]: DEBUG oslo_vmware.api [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 1033.496879] env[62692]: value = "task-1141667" [ 1033.496879] env[62692]: _type = "Task" [ 1033.496879] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.506240] env[62692]: DEBUG oslo_vmware.api [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141667, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.509355] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1033.509621] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1033.509825] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Deleting the datastore file [datastore2] 5a08fa8a-f9fe-4879-bb7b-baa04097df6c {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1033.510120] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eb01ef05-2a50-4bdf-9801-1ea7e3122170 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.516243] env[62692]: DEBUG oslo_vmware.api [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for the task: (returnval){ [ 1033.516243] env[62692]: value = "task-1141668" [ 1033.516243] env[62692]: _type = "Task" [ 1033.516243] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.525150] env[62692]: DEBUG oslo_vmware.api [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141668, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.581806] env[62692]: DEBUG oslo_vmware.api [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141662, 'name': PowerOnVM_Task, 'duration_secs': 0.806529} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.582814] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1033.582814] env[62692]: INFO nova.compute.manager [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Took 9.40 seconds to spawn the instance on the hypervisor. [ 1033.582814] env[62692]: DEBUG nova.compute.manager [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1033.583389] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-899dbcaf-6c71-4015-a5bb-35d0f3c33c34 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.810892] env[62692]: DEBUG oslo_concurrency.lockutils [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.317s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.813887] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.519s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.814161] env[62692]: DEBUG nova.objects.instance [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Lazy-loading 'resources' on Instance uuid f523c482-8365-47b8-9398-a56a98fcffe8 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.838176] env[62692]: DEBUG oslo_concurrency.lockutils [None req-25e44e32-4bcf-45a3-9769-eca012e35bf1 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Releasing lock "refresh_cache-d99737d8-2eb0-40ee-b61e-6c736c84ea59" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.839021] env[62692]: DEBUG nova.objects.instance [None req-25e44e32-4bcf-45a3-9769-eca012e35bf1 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lazy-loading 'migration_context' on Instance uuid d99737d8-2eb0-40ee-b61e-6c736c84ea59 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.839806] env[62692]: DEBUG oslo_vmware.api [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141653, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.007842] env[62692]: DEBUG oslo_vmware.api [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141667, 'name': PowerOffVM_Task, 'duration_secs': 0.252151} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.008210] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1034.008408] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1034.008670] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1c4f3f88-41d1-46fc-89b3-4c7d82ae1488 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.026556] env[62692]: DEBUG oslo_vmware.api [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Task: {'id': task-1141668, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177074} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.026811] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1034.027121] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1034.027406] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1034.027618] env[62692]: INFO nova.compute.manager [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Took 1.23 seconds to destroy the instance on the hypervisor. [ 1034.027898] env[62692]: DEBUG oslo.service.loopingcall [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1034.028180] env[62692]: DEBUG nova.compute.manager [-] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1034.028249] env[62692]: DEBUG nova.network.neutron [-] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1034.101061] env[62692]: INFO nova.compute.manager [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Took 30.78 seconds to build instance. [ 1034.323933] env[62692]: DEBUG oslo_concurrency.lockutils [None req-015253dc-65e5-41ed-966a-364c0069abb3 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "250b9cc2-706e-4fbf-887c-1c80fc27a0ed" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 34.426s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.328698] env[62692]: DEBUG oslo_concurrency.lockutils [None req-14d6ffc0-8145-4ba8-9f78-a28d05698ae8 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "250b9cc2-706e-4fbf-887c-1c80fc27a0ed" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 13.836s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.329008] env[62692]: DEBUG oslo_concurrency.lockutils [None req-14d6ffc0-8145-4ba8-9f78-a28d05698ae8 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "250b9cc2-706e-4fbf-887c-1c80fc27a0ed-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.329297] env[62692]: DEBUG oslo_concurrency.lockutils [None req-14d6ffc0-8145-4ba8-9f78-a28d05698ae8 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "250b9cc2-706e-4fbf-887c-1c80fc27a0ed-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.329348] env[62692]: DEBUG oslo_concurrency.lockutils [None req-14d6ffc0-8145-4ba8-9f78-a28d05698ae8 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "250b9cc2-706e-4fbf-887c-1c80fc27a0ed-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.338810] env[62692]: INFO nova.compute.manager [None req-14d6ffc0-8145-4ba8-9f78-a28d05698ae8 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Terminating instance [ 1034.340975] env[62692]: DEBUG nova.objects.base [None req-25e44e32-4bcf-45a3-9769-eca012e35bf1 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62692) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1034.342115] env[62692]: DEBUG nova.compute.manager [None req-14d6ffc0-8145-4ba8-9f78-a28d05698ae8 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1034.342115] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-14d6ffc0-8145-4ba8-9f78-a28d05698ae8 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1034.342868] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2ad4e5f-d689-4f21-8639-074d3d72d12a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.346469] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0b9e494c-35b7-4b22-aa24-625c848ad2f2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.352586] env[62692]: DEBUG oslo_vmware.api [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141653, 'name': ReconfigVM_Task, 'duration_secs': 5.759028} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.355837] env[62692]: DEBUG oslo_concurrency.lockutils [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.356479] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Reconfigured VM to detach interface {{(pid=62692) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1034.378339] env[62692]: DEBUG nova.compute.manager [req-6e2a0d4c-f99d-4622-9698-41c2885935b0 req-a409970e-38c2-4e2a-97b2-3f30ea96fc5e service nova] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Received event network-vif-deleted-6f540034-8dda-4def-adb2-0cdba3268510 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1034.378543] env[62692]: INFO nova.compute.manager [req-6e2a0d4c-f99d-4622-9698-41c2885935b0 req-a409970e-38c2-4e2a-97b2-3f30ea96fc5e service nova] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Neutron deleted interface 6f540034-8dda-4def-adb2-0cdba3268510; detaching it from the instance and deleting it from the info cache [ 1034.378761] env[62692]: DEBUG nova.network.neutron [req-6e2a0d4c-f99d-4622-9698-41c2885935b0 req-a409970e-38c2-4e2a-97b2-3f30ea96fc5e service nova] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.383449] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c8852b3-8af9-4ef5-bcf2-8fb3ee6e1e21 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.389795] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-231b85f2-a0ca-4a6d-8a56-5dce3ebafc17 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.408486] env[62692]: DEBUG oslo_vmware.api [None req-25e44e32-4bcf-45a3-9769-eca012e35bf1 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1034.408486] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52080de6-b1c1-1e1d-85da-d0c5dd600c02" [ 1034.408486] env[62692]: _type = "Task" [ 1034.408486] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.416489] env[62692]: DEBUG oslo_vmware.api [None req-25e44e32-4bcf-45a3-9769-eca012e35bf1 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52080de6-b1c1-1e1d-85da-d0c5dd600c02, 'name': SearchDatastore_Task, 'duration_secs': 0.007325} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.424951] env[62692]: DEBUG oslo_concurrency.lockutils [None req-25e44e32-4bcf-45a3-9769-eca012e35bf1 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.425879] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-14d6ffc0-8145-4ba8-9f78-a28d05698ae8 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 250b9cc2-706e-4fbf-887c-1c80fc27a0ed could not be found. [ 1034.426127] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-14d6ffc0-8145-4ba8-9f78-a28d05698ae8 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1034.426322] env[62692]: INFO nova.compute.manager [None req-14d6ffc0-8145-4ba8-9f78-a28d05698ae8 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Took 0.08 seconds to destroy the instance on the hypervisor. [ 1034.426687] env[62692]: DEBUG oslo.service.loopingcall [None req-14d6ffc0-8145-4ba8-9f78-a28d05698ae8 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1034.430956] env[62692]: DEBUG nova.compute.manager [-] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1034.431068] env[62692]: DEBUG nova.network.neutron [-] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1034.586859] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf959786-7d13-43ee-ba69-5639bf6abcd6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.594316] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e15a459a-b497-46db-bbde-b66995833345 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.623813] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b1c90024-15e5-48b2-8e8a-d47edaabac7d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "e8f645da-3b4a-4f21-9a17-aebdc8d112df" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.308s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.624862] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-defd3be4-d3eb-4cb4-823d-dcac2b6cab63 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.632733] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b19a87d4-08d1-4c37-b35f-c68388931bed {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.647841] env[62692]: DEBUG nova.compute.provider_tree [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1034.837058] env[62692]: DEBUG nova.network.neutron [-] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.888133] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8020c456-9a7b-48c8-b680-769ab4c7261f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.902402] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53fb015d-233f-499f-ba7d-263fee57e9c8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.935030] env[62692]: DEBUG nova.compute.manager [req-6e2a0d4c-f99d-4622-9698-41c2885935b0 req-a409970e-38c2-4e2a-97b2-3f30ea96fc5e service nova] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Detach interface failed, port_id=6f540034-8dda-4def-adb2-0cdba3268510, reason: Instance 5a08fa8a-f9fe-4879-bb7b-baa04097df6c could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1035.151358] env[62692]: DEBUG nova.scheduler.client.report [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1035.342756] env[62692]: INFO nova.compute.manager [-] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Took 1.31 seconds to deallocate network for instance. [ 1035.362186] env[62692]: DEBUG nova.network.neutron [-] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.656943] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.843s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.660050] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.727s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.660294] env[62692]: DEBUG nova.objects.instance [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Lazy-loading 'resources' on Instance uuid 16790505-cdf6-4937-8839-fa685a5f413e {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1035.686048] env[62692]: INFO nova.scheduler.client.report [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Deleted allocations for instance f523c482-8365-47b8-9398-a56a98fcffe8 [ 1035.854888] env[62692]: DEBUG oslo_concurrency.lockutils [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.865622] env[62692]: INFO nova.compute.manager [-] [instance: 250b9cc2-706e-4fbf-887c-1c80fc27a0ed] Took 1.43 seconds to deallocate network for instance. [ 1035.878568] env[62692]: DEBUG oslo_concurrency.lockutils [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "refresh_cache-668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.878763] env[62692]: DEBUG oslo_concurrency.lockutils [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "refresh_cache-668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.878945] env[62692]: DEBUG nova.network.neutron [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1036.002429] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "668fb0ea-7eed-4198-943c-8f916ec11368" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.006194] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "668fb0ea-7eed-4198-943c-8f916ec11368" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.007884] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "668fb0ea-7eed-4198-943c-8f916ec11368-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.007884] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "668fb0ea-7eed-4198-943c-8f916ec11368-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.007884] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "668fb0ea-7eed-4198-943c-8f916ec11368-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.015048] env[62692]: INFO nova.compute.manager [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Terminating instance [ 1036.016747] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1036.016960] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1036.017259] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Deleting the datastore file [datastore1] 754b3c65-1e4b-49d2-8980-095d975edb01 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1036.017576] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4013a1f5-1d2d-4373-b7f3-a86e3290db4e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.022182] env[62692]: DEBUG nova.compute.manager [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1036.022182] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1036.022182] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1036.022385] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1036.022496] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Deleting the datastore file [datastore1] ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1036.023690] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9891971f-9224-4e4f-9a06-545f8d666aaa {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.026484] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-83b768f3-131f-4a2a-8420-9294f641c7db {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.030405] env[62692]: DEBUG oslo_vmware.api [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1036.030405] env[62692]: value = "task-1141670" [ 1036.030405] env[62692]: _type = "Task" [ 1036.030405] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.038036] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1036.038036] env[62692]: DEBUG oslo_vmware.api [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 1036.038036] env[62692]: value = "task-1141671" [ 1036.038036] env[62692]: _type = "Task" [ 1036.038036] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.038413] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4c02eb66-7119-40aa-988e-5c4bb9ce1184 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.046732] env[62692]: DEBUG oslo_vmware.api [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141670, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.052325] env[62692]: DEBUG oslo_vmware.api [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141671, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.053734] env[62692]: DEBUG oslo_vmware.api [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1036.053734] env[62692]: value = "task-1141672" [ 1036.053734] env[62692]: _type = "Task" [ 1036.053734] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.063695] env[62692]: DEBUG oslo_vmware.api [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141672, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.194360] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5763320e-88ff-49ef-a545-c030425cf7b7 tempest-ServerExternalEventsTest-119880440 tempest-ServerExternalEventsTest-119880440-project-member] Lock "f523c482-8365-47b8-9398-a56a98fcffe8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.028s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.484446] env[62692]: DEBUG nova.compute.manager [req-5cee4abe-6825-4192-87eb-840645895468 req-cc820281-ec83-4089-8240-6caf6fa82aba service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Received event network-vif-deleted-e1a45cde-8d94-464d-8c3e-e3f94b62f9e6 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1036.484720] env[62692]: INFO nova.compute.manager [req-5cee4abe-6825-4192-87eb-840645895468 req-cc820281-ec83-4089-8240-6caf6fa82aba service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Neutron deleted interface e1a45cde-8d94-464d-8c3e-e3f94b62f9e6; detaching it from the instance and deleting it from the info cache [ 1036.485080] env[62692]: DEBUG nova.network.neutron [req-5cee4abe-6825-4192-87eb-840645895468 req-cc820281-ec83-4089-8240-6caf6fa82aba service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Updating instance_info_cache with network_info: [{"id": "2deb4259-2f64-4f7c-9289-eba76375335a", "address": "fa:16:3e:bd:3f:b3", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.135", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2deb4259-2f", "ovs_interfaceid": "2deb4259-2f64-4f7c-9289-eba76375335a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "dac9e9e1-d8ad-4ee7-8836-a84a7cb3cf27", "address": "fa:16:3e:69:e0:73", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdac9e9e1-d8", "ovs_interfaceid": "dac9e9e1-d8ad-4ee7-8836-a84a7cb3cf27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.499331] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f864cc23-6ac8-461b-be58-378e30b8efde {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.509074] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bab37e9-b893-4db3-b9ca-84fb69e4f4ad {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.557194] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99324937-d309-485f-9e7e-398142972a73 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.567096] env[62692]: DEBUG oslo_vmware.api [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141670, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.574131] env[62692]: DEBUG oslo_vmware.api [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141671, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.577358] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09f3b376-fa81-40c5-9896-854cd4d7f6de {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.586313] env[62692]: DEBUG oslo_vmware.api [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141672, 'name': PowerOffVM_Task, 'duration_secs': 0.241527} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.589587] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1036.590120] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1036.590215] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fbcab403-40f2-400c-b08d-a8e506bfd5cd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.600500] env[62692]: DEBUG nova.compute.provider_tree [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1036.697832] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1036.698091] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1036.698415] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Deleting the datastore file [datastore2] 668fb0ea-7eed-4198-943c-8f916ec11368 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1036.698544] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6667025e-0971-45cf-8e9d-e784c35a9af4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.704965] env[62692]: DEBUG oslo_vmware.api [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1036.704965] env[62692]: value = "task-1141674" [ 1036.704965] env[62692]: _type = "Task" [ 1036.704965] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.713996] env[62692]: DEBUG oslo_vmware.api [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141674, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.722804] env[62692]: INFO nova.network.neutron [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Port e1a45cde-8d94-464d-8c3e-e3f94b62f9e6 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1036.723015] env[62692]: INFO nova.network.neutron [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Port dac9e9e1-d8ad-4ee7-8836-a84a7cb3cf27 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1036.723481] env[62692]: DEBUG nova.network.neutron [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Updating instance_info_cache with network_info: [{"id": "2deb4259-2f64-4f7c-9289-eba76375335a", "address": "fa:16:3e:bd:3f:b3", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.135", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2deb4259-2f", "ovs_interfaceid": "2deb4259-2f64-4f7c-9289-eba76375335a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.894717] env[62692]: DEBUG oslo_concurrency.lockutils [None req-14d6ffc0-8145-4ba8-9f78-a28d05698ae8 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "250b9cc2-706e-4fbf-887c-1c80fc27a0ed" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.566s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.988247] env[62692]: DEBUG oslo_concurrency.lockutils [req-5cee4abe-6825-4192-87eb-840645895468 req-cc820281-ec83-4089-8240-6caf6fa82aba service nova] Acquiring lock "668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.061276] env[62692]: DEBUG oslo_vmware.api [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141671, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.861138} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.065057] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1037.065433] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1037.065747] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1037.066055] env[62692]: INFO nova.compute.manager [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Took 3.59 seconds to destroy the instance on the hypervisor. [ 1037.066454] env[62692]: DEBUG oslo.service.loopingcall [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1037.067616] env[62692]: DEBUG oslo_vmware.api [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141670, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.79374} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.067616] env[62692]: DEBUG nova.compute.manager [-] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1037.067616] env[62692]: DEBUG nova.network.neutron [-] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1037.069721] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1037.070161] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1037.070313] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1037.071107] env[62692]: INFO nova.compute.manager [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Took 4.53 seconds to destroy the instance on the hypervisor. [ 1037.071107] env[62692]: DEBUG oslo.service.loopingcall [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1037.071107] env[62692]: DEBUG nova.compute.manager [-] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1037.071107] env[62692]: DEBUG nova.network.neutron [-] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1037.104631] env[62692]: DEBUG nova.scheduler.client.report [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1037.217184] env[62692]: DEBUG oslo_vmware.api [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141674, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.185525} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.219482] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1037.219482] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1037.219482] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1037.219482] env[62692]: INFO nova.compute.manager [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Took 1.20 seconds to destroy the instance on the hypervisor. [ 1037.219482] env[62692]: DEBUG oslo.service.loopingcall [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1037.219482] env[62692]: DEBUG nova.compute.manager [-] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1037.219482] env[62692]: DEBUG nova.network.neutron [-] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1037.228023] env[62692]: DEBUG oslo_concurrency.lockutils [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "refresh_cache-668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.311779] env[62692]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port dac9e9e1-d8ad-4ee7-8836-a84a7cb3cf27 could not be found.", "detail": ""}} {{(pid=62692) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 1037.311779] env[62692]: DEBUG nova.network.neutron [-] Unable to show port dac9e9e1-d8ad-4ee7-8836-a84a7cb3cf27 as it no longer exists. {{(pid=62692) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 1037.614258] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.954s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.617581] env[62692]: DEBUG oslo_concurrency.lockutils [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.284s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.619508] env[62692]: INFO nova.compute.claims [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1037.651131] env[62692]: INFO nova.scheduler.client.report [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Deleted allocations for instance 16790505-cdf6-4937-8839-fa685a5f413e [ 1037.730549] env[62692]: DEBUG oslo_concurrency.lockutils [None req-429bee56-8b20-4bfa-9c0b-4dea9ed73648 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "interface-668fb0ea-7eed-4198-943c-8f916ec11368-e1a45cde-8d94-464d-8c3e-e3f94b62f9e6" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.994s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.071269] env[62692]: DEBUG nova.network.neutron [-] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.163561] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7884a119-2029-4caa-8cad-5b9690ca4c8a tempest-ServersTestFqdnHostnames-860710502 tempest-ServersTestFqdnHostnames-860710502-project-member] Lock "16790505-cdf6-4937-8839-fa685a5f413e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.369s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.310364] env[62692]: DEBUG oslo_concurrency.lockutils [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "ca56c2ec-95fc-44fc-856f-e395bda1703c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.310656] env[62692]: DEBUG oslo_concurrency.lockutils [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "ca56c2ec-95fc-44fc-856f-e395bda1703c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.313873] env[62692]: DEBUG nova.network.neutron [-] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.359185] env[62692]: DEBUG nova.compute.manager [req-82486a5e-c100-41ef-b228-3083ccf0d29b req-5be25d0a-9fa7-4a0e-892e-2b9bf4c25648 service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Received event network-vif-deleted-2deb4259-2f64-4f7c-9289-eba76375335a {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1038.359185] env[62692]: INFO nova.compute.manager [req-82486a5e-c100-41ef-b228-3083ccf0d29b req-5be25d0a-9fa7-4a0e-892e-2b9bf4c25648 service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Neutron deleted interface 2deb4259-2f64-4f7c-9289-eba76375335a; detaching it from the instance and deleting it from the info cache [ 1038.359510] env[62692]: DEBUG nova.network.neutron [req-82486a5e-c100-41ef-b228-3083ccf0d29b req-5be25d0a-9fa7-4a0e-892e-2b9bf4c25648 service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.535736] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d6bf423d-b277-4dd7-85bf-6118063e060f tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "7183ec4a-992d-48f1-8dda-7f499c2f4e1d" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.536167] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d6bf423d-b277-4dd7-85bf-6118063e060f tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "7183ec4a-992d-48f1-8dda-7f499c2f4e1d" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.575040] env[62692]: INFO nova.compute.manager [-] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Took 1.51 seconds to deallocate network for instance. [ 1038.576893] env[62692]: DEBUG nova.network.neutron [-] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.642010] env[62692]: DEBUG nova.compute.manager [req-aed9e422-14bf-497f-befa-a932e55318d4 req-094bcfab-3a90-402c-bf80-963fdbab2f94 service nova] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Received event network-vif-deleted-920fbe7b-7ba1-4d33-bf4f-d96609ed6640 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1038.642230] env[62692]: DEBUG nova.compute.manager [req-aed9e422-14bf-497f-befa-a932e55318d4 req-094bcfab-3a90-402c-bf80-963fdbab2f94 service nova] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Received event network-vif-deleted-152b902d-a99b-44d9-974f-71348e613fc9 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1038.815741] env[62692]: DEBUG nova.compute.manager [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1038.818996] env[62692]: INFO nova.compute.manager [-] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Took 1.75 seconds to deallocate network for instance. [ 1038.855636] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad893ecd-df14-478c-b3a1-89fbe543a28e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.864060] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-41cbda71-338a-4270-be48-ef5561ad6f51 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.867217] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adf8fe42-dbfc-4422-9724-0af8932c08b0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.903966] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-835b453f-837b-4ddb-afc0-6e4c656b6e05 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.915013] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29543bba-dd62-46ca-b8fb-ff50968284b7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.923886] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d324d66c-1a64-433b-9c57-346d5f143a29 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.935201] env[62692]: DEBUG nova.compute.manager [req-82486a5e-c100-41ef-b228-3083ccf0d29b req-5be25d0a-9fa7-4a0e-892e-2b9bf4c25648 service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Detach interface failed, port_id=2deb4259-2f64-4f7c-9289-eba76375335a, reason: Instance 668fb0ea-7eed-4198-943c-8f916ec11368 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1038.945778] env[62692]: DEBUG nova.compute.provider_tree [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1039.025690] env[62692]: DEBUG oslo_concurrency.lockutils [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "90932f05-fad3-4c6d-87ae-ab059351b0be" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.025979] env[62692]: DEBUG oslo_concurrency.lockutils [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "90932f05-fad3-4c6d-87ae-ab059351b0be" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.039640] env[62692]: DEBUG nova.compute.utils [None req-d6bf423d-b277-4dd7-85bf-6118063e060f tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1039.081926] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.082672] env[62692]: INFO nova.compute.manager [-] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Took 1.86 seconds to deallocate network for instance. [ 1039.215252] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Acquiring lock "e086d79d-29e6-41bf-a139-0b680ca0021d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.215506] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Lock "e086d79d-29e6-41bf-a139-0b680ca0021d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.328806] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.338979] env[62692]: DEBUG oslo_concurrency.lockutils [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.449295] env[62692]: DEBUG nova.scheduler.client.report [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1039.528342] env[62692]: DEBUG nova.compute.manager [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1039.544038] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d6bf423d-b277-4dd7-85bf-6118063e060f tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "7183ec4a-992d-48f1-8dda-7f499c2f4e1d" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.589305] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.718019] env[62692]: DEBUG nova.compute.manager [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1039.957931] env[62692]: DEBUG oslo_concurrency.lockutils [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.340s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.958454] env[62692]: DEBUG nova.compute.manager [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1039.962084] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.250s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.962216] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.965668] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.579s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.966381] env[62692]: INFO nova.compute.claims [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1039.991042] env[62692]: INFO nova.scheduler.client.report [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Deleted allocations for instance 3f6c9744-a6e3-43f3-8b6b-624ba178747d [ 1040.051867] env[62692]: DEBUG oslo_concurrency.lockutils [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.243523] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.472320] env[62692]: DEBUG nova.compute.utils [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1040.476544] env[62692]: DEBUG nova.compute.manager [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1040.476544] env[62692]: DEBUG nova.network.neutron [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1040.504223] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5616438a-2ff9-4c9f-86c5-95e1af4e1ed3 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "3f6c9744-a6e3-43f3-8b6b-624ba178747d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.669s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.525300] env[62692]: DEBUG nova.policy [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b40c8976b544c059458aa4e0dbab160', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9cb8ae08824245a7911de7455e7fcb68', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1040.613536] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d6bf423d-b277-4dd7-85bf-6118063e060f tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "7183ec4a-992d-48f1-8dda-7f499c2f4e1d" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.613817] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d6bf423d-b277-4dd7-85bf-6118063e060f tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "7183ec4a-992d-48f1-8dda-7f499c2f4e1d" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.614069] env[62692]: INFO nova.compute.manager [None req-d6bf423d-b277-4dd7-85bf-6118063e060f tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Attaching volume fc2d0515-f41f-4b2d-932b-af3a09d72939 to /dev/sdb [ 1040.655539] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17c848f5-47e6-456e-ae8a-7a35d426250c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.663763] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6f7154e-75c4-498a-8ab0-8ed72be6abed {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.677724] env[62692]: DEBUG nova.virt.block_device [None req-d6bf423d-b277-4dd7-85bf-6118063e060f tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Updating existing volume attachment record: 9047b94f-eff8-41c3-aef4-fab7265231e6 {{(pid=62692) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1040.844143] env[62692]: DEBUG nova.network.neutron [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Successfully created port: d5ff0422-e9b3-455c-bd28-422096794288 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1040.977099] env[62692]: DEBUG nova.compute.manager [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1041.220022] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0b97414-f2ff-4823-8f9e-22c233d63ea0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.227481] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-168d0972-f46d-4591-8fed-376ca63af68a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.257609] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbab30f4-56bf-47cc-a43e-fabc6feeb504 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.264800] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-250588e8-0040-49ad-b418-baed8ee79cb5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.278302] env[62692]: DEBUG nova.compute.provider_tree [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.781512] env[62692]: DEBUG nova.scheduler.client.report [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1041.992252] env[62692]: DEBUG nova.compute.manager [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1042.018129] env[62692]: DEBUG nova.virt.hardware [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1042.018420] env[62692]: DEBUG nova.virt.hardware [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1042.018585] env[62692]: DEBUG nova.virt.hardware [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1042.018770] env[62692]: DEBUG nova.virt.hardware [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1042.018923] env[62692]: DEBUG nova.virt.hardware [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1042.019090] env[62692]: DEBUG nova.virt.hardware [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1042.019340] env[62692]: DEBUG nova.virt.hardware [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1042.019509] env[62692]: DEBUG nova.virt.hardware [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1042.019703] env[62692]: DEBUG nova.virt.hardware [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1042.019942] env[62692]: DEBUG nova.virt.hardware [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1042.020172] env[62692]: DEBUG nova.virt.hardware [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1042.021100] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaac45cf-3a94-4e62-abe1-c5681e52441f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.029348] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8248b36-df78-44ee-b5e0-e15b1d9c63e0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.284619] env[62692]: DEBUG nova.compute.manager [req-aaeb0454-5bb7-4e2d-ac80-7523a4934870 req-fbfcb6b4-f380-4912-b096-499b04baf4dc service nova] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Received event network-vif-plugged-d5ff0422-e9b3-455c-bd28-422096794288 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1042.284932] env[62692]: DEBUG oslo_concurrency.lockutils [req-aaeb0454-5bb7-4e2d-ac80-7523a4934870 req-fbfcb6b4-f380-4912-b096-499b04baf4dc service nova] Acquiring lock "9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.285345] env[62692]: DEBUG oslo_concurrency.lockutils [req-aaeb0454-5bb7-4e2d-ac80-7523a4934870 req-fbfcb6b4-f380-4912-b096-499b04baf4dc service nova] Lock "9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.285620] env[62692]: DEBUG oslo_concurrency.lockutils [req-aaeb0454-5bb7-4e2d-ac80-7523a4934870 req-fbfcb6b4-f380-4912-b096-499b04baf4dc service nova] Lock "9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.285885] env[62692]: DEBUG nova.compute.manager [req-aaeb0454-5bb7-4e2d-ac80-7523a4934870 req-fbfcb6b4-f380-4912-b096-499b04baf4dc service nova] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] No waiting events found dispatching network-vif-plugged-d5ff0422-e9b3-455c-bd28-422096794288 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1042.286173] env[62692]: WARNING nova.compute.manager [req-aaeb0454-5bb7-4e2d-ac80-7523a4934870 req-fbfcb6b4-f380-4912-b096-499b04baf4dc service nova] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Received unexpected event network-vif-plugged-d5ff0422-e9b3-455c-bd28-422096794288 for instance with vm_state building and task_state spawning. [ 1042.287247] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.323s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.287782] env[62692]: DEBUG nova.compute.manager [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1042.290681] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.166s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.293072] env[62692]: INFO nova.compute.claims [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1042.380407] env[62692]: DEBUG nova.network.neutron [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Successfully updated port: d5ff0422-e9b3-455c-bd28-422096794288 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1042.800498] env[62692]: DEBUG nova.compute.utils [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1042.801770] env[62692]: DEBUG nova.compute.manager [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1042.801877] env[62692]: DEBUG nova.network.neutron [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1042.852518] env[62692]: DEBUG nova.policy [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5f4f8dca6b6549a7b3aac06e3e64ba08', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f26b93a4a43b45e8a11f3793996c8496', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1042.883113] env[62692]: DEBUG oslo_concurrency.lockutils [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "refresh_cache-9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.883276] env[62692]: DEBUG oslo_concurrency.lockutils [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired lock "refresh_cache-9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.883414] env[62692]: DEBUG nova.network.neutron [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1043.109217] env[62692]: DEBUG nova.network.neutron [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Successfully created port: 0b794a90-367f-4d18-a270-df14f204b2a3 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1043.307349] env[62692]: DEBUG nova.compute.manager [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1043.421327] env[62692]: DEBUG nova.network.neutron [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1043.537032] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87afa8bd-a300-4f14-91cc-c2f6f3d2d909 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.544739] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-128c87ac-cdb1-4369-afc2-4d669871d9a6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.577277] env[62692]: DEBUG nova.network.neutron [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Updating instance_info_cache with network_info: [{"id": "d5ff0422-e9b3-455c-bd28-422096794288", "address": "fa:16:3e:46:38:ff", "network": {"id": "f6352999-f943-431d-b54a-c9dc1d810a85", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-121037111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9cb8ae08824245a7911de7455e7fcb68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c842425c-544e-4ce2-9657-512723bd318e", "external-id": "nsx-vlan-transportzone-80", "segmentation_id": 80, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5ff0422-e9", "ovs_interfaceid": "d5ff0422-e9b3-455c-bd28-422096794288", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.578272] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dd0c4f1-3f52-4c5b-b7bc-f8f187d4abc7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.586766] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27e7bd71-f9fe-4cea-8ada-5e9ccfe06d9a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.601367] env[62692]: DEBUG nova.compute.provider_tree [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1044.082023] env[62692]: DEBUG oslo_concurrency.lockutils [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Releasing lock "refresh_cache-9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.082385] env[62692]: DEBUG nova.compute.manager [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Instance network_info: |[{"id": "d5ff0422-e9b3-455c-bd28-422096794288", "address": "fa:16:3e:46:38:ff", "network": {"id": "f6352999-f943-431d-b54a-c9dc1d810a85", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-121037111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9cb8ae08824245a7911de7455e7fcb68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c842425c-544e-4ce2-9657-512723bd318e", "external-id": "nsx-vlan-transportzone-80", "segmentation_id": 80, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5ff0422-e9", "ovs_interfaceid": "d5ff0422-e9b3-455c-bd28-422096794288", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1044.083317] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:46:38:ff', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c842425c-544e-4ce2-9657-512723bd318e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd5ff0422-e9b3-455c-bd28-422096794288', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1044.090575] env[62692]: DEBUG oslo.service.loopingcall [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1044.090858] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1044.091683] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-81fbad35-5a6f-47de-933f-e119636fdc25 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.107492] env[62692]: DEBUG nova.scheduler.client.report [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1044.116010] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1044.116010] env[62692]: value = "task-1141679" [ 1044.116010] env[62692]: _type = "Task" [ 1044.116010] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.125899] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141679, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.309750] env[62692]: DEBUG nova.compute.manager [req-6362dfe9-7cbd-4ea6-bca2-6ec4ad6727e6 req-fddfa9d8-3440-4eab-b0ff-8e8c22d6b01b service nova] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Received event network-changed-d5ff0422-e9b3-455c-bd28-422096794288 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1044.310060] env[62692]: DEBUG nova.compute.manager [req-6362dfe9-7cbd-4ea6-bca2-6ec4ad6727e6 req-fddfa9d8-3440-4eab-b0ff-8e8c22d6b01b service nova] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Refreshing instance network info cache due to event network-changed-d5ff0422-e9b3-455c-bd28-422096794288. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1044.310304] env[62692]: DEBUG oslo_concurrency.lockutils [req-6362dfe9-7cbd-4ea6-bca2-6ec4ad6727e6 req-fddfa9d8-3440-4eab-b0ff-8e8c22d6b01b service nova] Acquiring lock "refresh_cache-9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.310458] env[62692]: DEBUG oslo_concurrency.lockutils [req-6362dfe9-7cbd-4ea6-bca2-6ec4ad6727e6 req-fddfa9d8-3440-4eab-b0ff-8e8c22d6b01b service nova] Acquired lock "refresh_cache-9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.310625] env[62692]: DEBUG nova.network.neutron [req-6362dfe9-7cbd-4ea6-bca2-6ec4ad6727e6 req-fddfa9d8-3440-4eab-b0ff-8e8c22d6b01b service nova] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Refreshing network info cache for port d5ff0422-e9b3-455c-bd28-422096794288 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1044.317276] env[62692]: DEBUG nova.compute.manager [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1044.345477] env[62692]: DEBUG nova.virt.hardware [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1044.345781] env[62692]: DEBUG nova.virt.hardware [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1044.346041] env[62692]: DEBUG nova.virt.hardware [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1044.346296] env[62692]: DEBUG nova.virt.hardware [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1044.346555] env[62692]: DEBUG nova.virt.hardware [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1044.346796] env[62692]: DEBUG nova.virt.hardware [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1044.347083] env[62692]: DEBUG nova.virt.hardware [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1044.347302] env[62692]: DEBUG nova.virt.hardware [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1044.347529] env[62692]: DEBUG nova.virt.hardware [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1044.347754] env[62692]: DEBUG nova.virt.hardware [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1044.347976] env[62692]: DEBUG nova.virt.hardware [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1044.349160] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7168d866-e978-493f-b7d7-795a7bad6e1b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.357406] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f18ecea-d108-4bca-a76f-c8e28012bc92 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.612667] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.322s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.613362] env[62692]: DEBUG nova.compute.manager [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1044.619985] env[62692]: DEBUG nova.network.neutron [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Successfully updated port: 0b794a90-367f-4d18-a270-df14f204b2a3 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1044.620988] env[62692]: DEBUG oslo_concurrency.lockutils [None req-25e44e32-4bcf-45a3-9769-eca012e35bf1 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 10.196s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.631734] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141679, 'name': CreateVM_Task, 'duration_secs': 0.307377} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.632469] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1044.633190] env[62692]: DEBUG oslo_concurrency.lockutils [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.633366] env[62692]: DEBUG oslo_concurrency.lockutils [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.633710] env[62692]: DEBUG oslo_concurrency.lockutils [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1044.634235] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66f15bb8-0f4c-453e-922f-6d61ec695451 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.640052] env[62692]: DEBUG oslo_vmware.api [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1044.640052] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]526d60d0-d987-d0e4-e1f1-e96eebbcc0f2" [ 1044.640052] env[62692]: _type = "Task" [ 1044.640052] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.649645] env[62692]: DEBUG oslo_vmware.api [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]526d60d0-d987-d0e4-e1f1-e96eebbcc0f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.120852] env[62692]: DEBUG nova.compute.utils [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1045.122394] env[62692]: DEBUG nova.compute.manager [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1045.123444] env[62692]: DEBUG nova.network.neutron [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1045.126694] env[62692]: DEBUG nova.network.neutron [req-6362dfe9-7cbd-4ea6-bca2-6ec4ad6727e6 req-fddfa9d8-3440-4eab-b0ff-8e8c22d6b01b service nova] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Updated VIF entry in instance network info cache for port d5ff0422-e9b3-455c-bd28-422096794288. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1045.126694] env[62692]: DEBUG nova.network.neutron [req-6362dfe9-7cbd-4ea6-bca2-6ec4ad6727e6 req-fddfa9d8-3440-4eab-b0ff-8e8c22d6b01b service nova] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Updating instance_info_cache with network_info: [{"id": "d5ff0422-e9b3-455c-bd28-422096794288", "address": "fa:16:3e:46:38:ff", "network": {"id": "f6352999-f943-431d-b54a-c9dc1d810a85", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-121037111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9cb8ae08824245a7911de7455e7fcb68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c842425c-544e-4ce2-9657-512723bd318e", "external-id": "nsx-vlan-transportzone-80", "segmentation_id": 80, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5ff0422-e9", "ovs_interfaceid": "d5ff0422-e9b3-455c-bd28-422096794288", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.126821] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquiring lock "refresh_cache-af260f9a-4e9b-4af1-90da-c619c4755eca" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.126868] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquired lock "refresh_cache-af260f9a-4e9b-4af1-90da-c619c4755eca" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.127545] env[62692]: DEBUG nova.network.neutron [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1045.151229] env[62692]: DEBUG oslo_vmware.api [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]526d60d0-d987-d0e4-e1f1-e96eebbcc0f2, 'name': SearchDatastore_Task, 'duration_secs': 0.018445} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.153794] env[62692]: DEBUG oslo_concurrency.lockutils [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.154089] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1045.154341] env[62692]: DEBUG oslo_concurrency.lockutils [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.154495] env[62692]: DEBUG oslo_concurrency.lockutils [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.154677] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1045.155888] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-70095edf-a68b-40bb-bf0b-5cacecd1d17f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.163853] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1045.164115] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1045.164886] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67b843e8-4462-4292-b322-8278817fea0f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.172674] env[62692]: DEBUG oslo_vmware.api [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1045.172674] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]527b64cf-47f0-4849-bf32-676b349c3670" [ 1045.172674] env[62692]: _type = "Task" [ 1045.172674] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.180402] env[62692]: DEBUG oslo_vmware.api [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]527b64cf-47f0-4849-bf32-676b349c3670, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.183832] env[62692]: DEBUG nova.policy [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f2e2ee26163841ddb15bb67520e98394', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e13f8888d5e84eed8247101558e29a56', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1045.231065] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6bf423d-b277-4dd7-85bf-6118063e060f tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Volume attach. Driver type: vmdk {{(pid=62692) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1045.231333] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6bf423d-b277-4dd7-85bf-6118063e060f tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249028', 'volume_id': 'fc2d0515-f41f-4b2d-932b-af3a09d72939', 'name': 'volume-fc2d0515-f41f-4b2d-932b-af3a09d72939', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7183ec4a-992d-48f1-8dda-7f499c2f4e1d', 'attached_at': '', 'detached_at': '', 'volume_id': 'fc2d0515-f41f-4b2d-932b-af3a09d72939', 'serial': 'fc2d0515-f41f-4b2d-932b-af3a09d72939'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1045.232250] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5efd1a30-1112-4519-bd4a-2d5ed4ea4cfa {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.253929] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdf70d15-6556-45d0-a85b-28057d0cb4f7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.278768] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6bf423d-b277-4dd7-85bf-6118063e060f tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] volume-fc2d0515-f41f-4b2d-932b-af3a09d72939/volume-fc2d0515-f41f-4b2d-932b-af3a09d72939.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1045.281654] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d7c4230c-733f-47ea-983f-cb84f28b2603 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.301937] env[62692]: DEBUG oslo_vmware.api [None req-d6bf423d-b277-4dd7-85bf-6118063e060f tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1045.301937] env[62692]: value = "task-1141680" [ 1045.301937] env[62692]: _type = "Task" [ 1045.301937] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.308896] env[62692]: DEBUG oslo_vmware.api [None req-d6bf423d-b277-4dd7-85bf-6118063e060f tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141680, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.417838] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b064d5c5-ccab-45e4-b4e5-122b3f70ebb6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.426696] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-184bc0f7-b1a3-450a-bb61-8da914b1a373 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.464256] env[62692]: DEBUG nova.network.neutron [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Successfully created port: 5427dd91-588d-4a85-9ca4-951c8a2588be {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1045.466911] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-321fbc9f-2c0f-415b-9c87-79d0a51fd05c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.474856] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fdcf446-a480-4f33-a5ed-eb286a9da55b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.497100] env[62692]: DEBUG nova.compute.provider_tree [None req-25e44e32-4bcf-45a3-9769-eca012e35bf1 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1045.625980] env[62692]: DEBUG nova.compute.manager [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1045.634052] env[62692]: DEBUG oslo_concurrency.lockutils [req-6362dfe9-7cbd-4ea6-bca2-6ec4ad6727e6 req-fddfa9d8-3440-4eab-b0ff-8e8c22d6b01b service nova] Releasing lock "refresh_cache-9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.664588] env[62692]: DEBUG nova.network.neutron [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1045.688928] env[62692]: DEBUG oslo_vmware.api [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]527b64cf-47f0-4849-bf32-676b349c3670, 'name': SearchDatastore_Task, 'duration_secs': 0.008365} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.692665] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86c2b860-ad49-404e-b0d0-2002c24e41e6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.700476] env[62692]: DEBUG oslo_vmware.api [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1045.700476] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52bbc879-4a2c-c0b3-69b2-90cb7c4ddfaa" [ 1045.700476] env[62692]: _type = "Task" [ 1045.700476] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.709722] env[62692]: DEBUG oslo_vmware.api [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52bbc879-4a2c-c0b3-69b2-90cb7c4ddfaa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.815117] env[62692]: DEBUG oslo_vmware.api [None req-d6bf423d-b277-4dd7-85bf-6118063e060f tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141680, 'name': ReconfigVM_Task, 'duration_secs': 0.399718} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.815543] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6bf423d-b277-4dd7-85bf-6118063e060f tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Reconfigured VM instance instance-00000055 to attach disk [datastore2] volume-fc2d0515-f41f-4b2d-932b-af3a09d72939/volume-fc2d0515-f41f-4b2d-932b-af3a09d72939.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1045.823577] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3ef3db35-25f1-406d-9242-17db3acd2cda {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.840943] env[62692]: DEBUG oslo_vmware.api [None req-d6bf423d-b277-4dd7-85bf-6118063e060f tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1045.840943] env[62692]: value = "task-1141681" [ 1045.840943] env[62692]: _type = "Task" [ 1045.840943] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.844817] env[62692]: DEBUG nova.network.neutron [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Updating instance_info_cache with network_info: [{"id": "0b794a90-367f-4d18-a270-df14f204b2a3", "address": "fa:16:3e:45:cd:d0", "network": {"id": "b7c23ada-5498-4686-bd7b-b32d0837e401", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1234309205-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f26b93a4a43b45e8a11f3793996c8496", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b67e519-46cf-44ce-b670-4ba4c0c5b658", "external-id": "nsx-vlan-transportzone-110", "segmentation_id": 110, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b794a90-36", "ovs_interfaceid": "0b794a90-367f-4d18-a270-df14f204b2a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.852400] env[62692]: DEBUG oslo_vmware.api [None req-d6bf423d-b277-4dd7-85bf-6118063e060f tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141681, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.998701] env[62692]: DEBUG nova.scheduler.client.report [None req-25e44e32-4bcf-45a3-9769-eca012e35bf1 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1046.213549] env[62692]: DEBUG oslo_vmware.api [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52bbc879-4a2c-c0b3-69b2-90cb7c4ddfaa, 'name': SearchDatastore_Task, 'duration_secs': 0.011331} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.213873] env[62692]: DEBUG oslo_concurrency.lockutils [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.214201] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da/9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1046.214510] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-443137c3-5031-42c3-a65e-665385fbfa6d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.222610] env[62692]: DEBUG oslo_vmware.api [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1046.222610] env[62692]: value = "task-1141682" [ 1046.222610] env[62692]: _type = "Task" [ 1046.222610] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.230449] env[62692]: DEBUG oslo_vmware.api [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141682, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.336191] env[62692]: DEBUG nova.compute.manager [req-2af66a42-2198-4458-b049-cd8ec4dfe5c2 req-a6100108-2764-4cf5-930e-46af7a5aeba5 service nova] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Received event network-vif-plugged-0b794a90-367f-4d18-a270-df14f204b2a3 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1046.336551] env[62692]: DEBUG oslo_concurrency.lockutils [req-2af66a42-2198-4458-b049-cd8ec4dfe5c2 req-a6100108-2764-4cf5-930e-46af7a5aeba5 service nova] Acquiring lock "af260f9a-4e9b-4af1-90da-c619c4755eca-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.336898] env[62692]: DEBUG oslo_concurrency.lockutils [req-2af66a42-2198-4458-b049-cd8ec4dfe5c2 req-a6100108-2764-4cf5-930e-46af7a5aeba5 service nova] Lock "af260f9a-4e9b-4af1-90da-c619c4755eca-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.337209] env[62692]: DEBUG oslo_concurrency.lockutils [req-2af66a42-2198-4458-b049-cd8ec4dfe5c2 req-a6100108-2764-4cf5-930e-46af7a5aeba5 service nova] Lock "af260f9a-4e9b-4af1-90da-c619c4755eca-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.337504] env[62692]: DEBUG nova.compute.manager [req-2af66a42-2198-4458-b049-cd8ec4dfe5c2 req-a6100108-2764-4cf5-930e-46af7a5aeba5 service nova] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] No waiting events found dispatching network-vif-plugged-0b794a90-367f-4d18-a270-df14f204b2a3 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1046.337793] env[62692]: WARNING nova.compute.manager [req-2af66a42-2198-4458-b049-cd8ec4dfe5c2 req-a6100108-2764-4cf5-930e-46af7a5aeba5 service nova] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Received unexpected event network-vif-plugged-0b794a90-367f-4d18-a270-df14f204b2a3 for instance with vm_state building and task_state spawning. [ 1046.338102] env[62692]: DEBUG nova.compute.manager [req-2af66a42-2198-4458-b049-cd8ec4dfe5c2 req-a6100108-2764-4cf5-930e-46af7a5aeba5 service nova] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Received event network-changed-0b794a90-367f-4d18-a270-df14f204b2a3 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1046.338399] env[62692]: DEBUG nova.compute.manager [req-2af66a42-2198-4458-b049-cd8ec4dfe5c2 req-a6100108-2764-4cf5-930e-46af7a5aeba5 service nova] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Refreshing instance network info cache due to event network-changed-0b794a90-367f-4d18-a270-df14f204b2a3. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1046.338698] env[62692]: DEBUG oslo_concurrency.lockutils [req-2af66a42-2198-4458-b049-cd8ec4dfe5c2 req-a6100108-2764-4cf5-930e-46af7a5aeba5 service nova] Acquiring lock "refresh_cache-af260f9a-4e9b-4af1-90da-c619c4755eca" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.347220] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Releasing lock "refresh_cache-af260f9a-4e9b-4af1-90da-c619c4755eca" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.347532] env[62692]: DEBUG nova.compute.manager [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Instance network_info: |[{"id": "0b794a90-367f-4d18-a270-df14f204b2a3", "address": "fa:16:3e:45:cd:d0", "network": {"id": "b7c23ada-5498-4686-bd7b-b32d0837e401", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1234309205-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f26b93a4a43b45e8a11f3793996c8496", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b67e519-46cf-44ce-b670-4ba4c0c5b658", "external-id": "nsx-vlan-transportzone-110", "segmentation_id": 110, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b794a90-36", "ovs_interfaceid": "0b794a90-367f-4d18-a270-df14f204b2a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1046.348182] env[62692]: DEBUG oslo_concurrency.lockutils [req-2af66a42-2198-4458-b049-cd8ec4dfe5c2 req-a6100108-2764-4cf5-930e-46af7a5aeba5 service nova] Acquired lock "refresh_cache-af260f9a-4e9b-4af1-90da-c619c4755eca" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.348182] env[62692]: DEBUG nova.network.neutron [req-2af66a42-2198-4458-b049-cd8ec4dfe5c2 req-a6100108-2764-4cf5-930e-46af7a5aeba5 service nova] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Refreshing network info cache for port 0b794a90-367f-4d18-a270-df14f204b2a3 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1046.349979] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:45:cd:d0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3b67e519-46cf-44ce-b670-4ba4c0c5b658', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0b794a90-367f-4d18-a270-df14f204b2a3', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1046.357946] env[62692]: DEBUG oslo.service.loopingcall [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1046.362715] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1046.363934] env[62692]: DEBUG oslo_vmware.api [None req-d6bf423d-b277-4dd7-85bf-6118063e060f tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141681, 'name': ReconfigVM_Task, 'duration_secs': 0.131007} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.363934] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-09f2cca8-cada-480a-a4b0-4af7dc4f4b06 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.379872] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6bf423d-b277-4dd7-85bf-6118063e060f tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249028', 'volume_id': 'fc2d0515-f41f-4b2d-932b-af3a09d72939', 'name': 'volume-fc2d0515-f41f-4b2d-932b-af3a09d72939', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7183ec4a-992d-48f1-8dda-7f499c2f4e1d', 'attached_at': '', 'detached_at': '', 'volume_id': 'fc2d0515-f41f-4b2d-932b-af3a09d72939', 'serial': 'fc2d0515-f41f-4b2d-932b-af3a09d72939'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1046.388735] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1046.388735] env[62692]: value = "task-1141683" [ 1046.388735] env[62692]: _type = "Task" [ 1046.388735] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.397464] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141683, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.635783] env[62692]: DEBUG nova.compute.manager [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1046.664699] env[62692]: DEBUG nova.virt.hardware [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1046.664967] env[62692]: DEBUG nova.virt.hardware [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1046.665150] env[62692]: DEBUG nova.virt.hardware [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1046.665343] env[62692]: DEBUG nova.virt.hardware [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1046.665501] env[62692]: DEBUG nova.virt.hardware [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1046.665654] env[62692]: DEBUG nova.virt.hardware [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1046.665869] env[62692]: DEBUG nova.virt.hardware [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1046.666047] env[62692]: DEBUG nova.virt.hardware [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1046.666230] env[62692]: DEBUG nova.virt.hardware [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1046.666400] env[62692]: DEBUG nova.virt.hardware [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1046.666581] env[62692]: DEBUG nova.virt.hardware [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1046.667608] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fccb74c5-754a-4604-a5b8-016a8c47357c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.675478] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8069e0c-f33c-4ba3-aadf-21959d613feb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.732477] env[62692]: DEBUG oslo_vmware.api [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141682, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.464387} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.732899] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da/9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1046.733252] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1046.733614] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2fd3de4a-1b86-494a-823e-4aaab07bb55b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.740629] env[62692]: DEBUG oslo_vmware.api [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1046.740629] env[62692]: value = "task-1141684" [ 1046.740629] env[62692]: _type = "Task" [ 1046.740629] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.747914] env[62692]: DEBUG oslo_vmware.api [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141684, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.900526] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141683, 'name': CreateVM_Task, 'duration_secs': 0.348363} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.900701] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1046.901455] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.901583] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.901926] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1046.902183] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-325289fc-aacd-455f-aeca-a0aab48d9507 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.906823] env[62692]: DEBUG oslo_vmware.api [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Waiting for the task: (returnval){ [ 1046.906823] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52f98a04-4054-4389-44a7-36690ec664ed" [ 1046.906823] env[62692]: _type = "Task" [ 1046.906823] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.917331] env[62692]: DEBUG oslo_vmware.api [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52f98a04-4054-4389-44a7-36690ec664ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.009491] env[62692]: DEBUG oslo_concurrency.lockutils [None req-25e44e32-4bcf-45a3-9769-eca012e35bf1 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.388s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.013125] env[62692]: DEBUG oslo_concurrency.lockutils [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.158s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.013403] env[62692]: DEBUG nova.objects.instance [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lazy-loading 'resources' on Instance uuid 5a08fa8a-f9fe-4879-bb7b-baa04097df6c {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1047.138504] env[62692]: DEBUG nova.network.neutron [req-2af66a42-2198-4458-b049-cd8ec4dfe5c2 req-a6100108-2764-4cf5-930e-46af7a5aeba5 service nova] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Updated VIF entry in instance network info cache for port 0b794a90-367f-4d18-a270-df14f204b2a3. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1047.138877] env[62692]: DEBUG nova.network.neutron [req-2af66a42-2198-4458-b049-cd8ec4dfe5c2 req-a6100108-2764-4cf5-930e-46af7a5aeba5 service nova] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Updating instance_info_cache with network_info: [{"id": "0b794a90-367f-4d18-a270-df14f204b2a3", "address": "fa:16:3e:45:cd:d0", "network": {"id": "b7c23ada-5498-4686-bd7b-b32d0837e401", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1234309205-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f26b93a4a43b45e8a11f3793996c8496", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b67e519-46cf-44ce-b670-4ba4c0c5b658", "external-id": "nsx-vlan-transportzone-110", "segmentation_id": 110, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b794a90-36", "ovs_interfaceid": "0b794a90-367f-4d18-a270-df14f204b2a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.250241] env[62692]: DEBUG oslo_vmware.api [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141684, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066079} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.250577] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1047.251351] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e13ae8e-9ed9-4ef7-aafa-34fb4e7d889c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.275224] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da/9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1047.275539] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-713f81fd-968d-41a1-aa66-68ead4b49ba7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.295910] env[62692]: DEBUG oslo_vmware.api [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1047.295910] env[62692]: value = "task-1141685" [ 1047.295910] env[62692]: _type = "Task" [ 1047.295910] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.305209] env[62692]: DEBUG oslo_vmware.api [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141685, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.418367] env[62692]: DEBUG oslo_vmware.api [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52f98a04-4054-4389-44a7-36690ec664ed, 'name': SearchDatastore_Task, 'duration_secs': 0.011521} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.418730] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.418996] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1047.419279] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.419444] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.419636] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1047.420767] env[62692]: DEBUG nova.objects.instance [None req-d6bf423d-b277-4dd7-85bf-6118063e060f tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lazy-loading 'flavor' on Instance uuid 7183ec4a-992d-48f1-8dda-7f499c2f4e1d {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1047.422126] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aa15f06c-b8bb-43b5-8ca6-9f7b23f9f36f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.430634] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1047.430825] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1047.431573] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1362d4af-d78f-4a65-9926-10d461701b54 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.436820] env[62692]: DEBUG oslo_vmware.api [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Waiting for the task: (returnval){ [ 1047.436820] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]525a2162-c439-0a16-cbb5-e4c688913e7a" [ 1047.436820] env[62692]: _type = "Task" [ 1047.436820] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.446022] env[62692]: DEBUG oslo_vmware.api [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]525a2162-c439-0a16-cbb5-e4c688913e7a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.578245] env[62692]: INFO nova.scheduler.client.report [None req-25e44e32-4bcf-45a3-9769-eca012e35bf1 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Deleted allocation for migration 91b03d62-ef95-4521-8e30-c12febaeacc3 [ 1047.607103] env[62692]: DEBUG nova.compute.manager [req-5809ff43-e825-41f8-85a1-c6791c567113 req-51081339-5dd1-4692-855c-ea1caa807310 service nova] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Received event network-vif-plugged-5427dd91-588d-4a85-9ca4-951c8a2588be {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1047.607310] env[62692]: DEBUG oslo_concurrency.lockutils [req-5809ff43-e825-41f8-85a1-c6791c567113 req-51081339-5dd1-4692-855c-ea1caa807310 service nova] Acquiring lock "a1951dba-ec5a-4849-9e22-37fba08ec60c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.607547] env[62692]: DEBUG oslo_concurrency.lockutils [req-5809ff43-e825-41f8-85a1-c6791c567113 req-51081339-5dd1-4692-855c-ea1caa807310 service nova] Lock "a1951dba-ec5a-4849-9e22-37fba08ec60c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.607687] env[62692]: DEBUG oslo_concurrency.lockutils [req-5809ff43-e825-41f8-85a1-c6791c567113 req-51081339-5dd1-4692-855c-ea1caa807310 service nova] Lock "a1951dba-ec5a-4849-9e22-37fba08ec60c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.607948] env[62692]: DEBUG nova.compute.manager [req-5809ff43-e825-41f8-85a1-c6791c567113 req-51081339-5dd1-4692-855c-ea1caa807310 service nova] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] No waiting events found dispatching network-vif-plugged-5427dd91-588d-4a85-9ca4-951c8a2588be {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1047.608071] env[62692]: WARNING nova.compute.manager [req-5809ff43-e825-41f8-85a1-c6791c567113 req-51081339-5dd1-4692-855c-ea1caa807310 service nova] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Received unexpected event network-vif-plugged-5427dd91-588d-4a85-9ca4-951c8a2588be for instance with vm_state building and task_state spawning. [ 1047.641384] env[62692]: DEBUG oslo_concurrency.lockutils [req-2af66a42-2198-4458-b049-cd8ec4dfe5c2 req-a6100108-2764-4cf5-930e-46af7a5aeba5 service nova] Releasing lock "refresh_cache-af260f9a-4e9b-4af1-90da-c619c4755eca" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.646810] env[62692]: DEBUG nova.network.neutron [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Successfully updated port: 5427dd91-588d-4a85-9ca4-951c8a2588be {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1047.789681] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb297d31-4451-4163-bac9-43edbb7d9640 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.800796] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-294b89af-347a-4323-a8a7-f4257012ef72 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.812034] env[62692]: DEBUG oslo_vmware.api [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141685, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.843259] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b67d45-df08-4cdc-a99c-7c987811718c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.851542] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8787d510-ac8b-4d85-a986-f8abc175058c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.868415] env[62692]: DEBUG nova.compute.provider_tree [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1047.927594] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d6bf423d-b277-4dd7-85bf-6118063e060f tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "7183ec4a-992d-48f1-8dda-7f499c2f4e1d" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.314s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.948163] env[62692]: DEBUG oslo_vmware.api [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]525a2162-c439-0a16-cbb5-e4c688913e7a, 'name': SearchDatastore_Task, 'duration_secs': 0.008763} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.949132] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5fd8ec67-39cd-4e6b-8dac-88fc88a136b4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.954481] env[62692]: DEBUG oslo_vmware.api [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Waiting for the task: (returnval){ [ 1047.954481] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]529f4465-edd2-3ee9-aaee-e8d3da948bfc" [ 1047.954481] env[62692]: _type = "Task" [ 1047.954481] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.963913] env[62692]: DEBUG oslo_vmware.api [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]529f4465-edd2-3ee9-aaee-e8d3da948bfc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.011125] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6a49e111-9fe9-4008-a52a-349e6f75090d tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "7183ec4a-992d-48f1-8dda-7f499c2f4e1d" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.011476] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6a49e111-9fe9-4008-a52a-349e6f75090d tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "7183ec4a-992d-48f1-8dda-7f499c2f4e1d" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.011697] env[62692]: DEBUG nova.compute.manager [None req-6a49e111-9fe9-4008-a52a-349e6f75090d tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1048.012661] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc15fe2d-506d-4c79-965f-c52a94791a20 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.020373] env[62692]: DEBUG nova.compute.manager [None req-6a49e111-9fe9-4008-a52a-349e6f75090d tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62692) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1048.021027] env[62692]: DEBUG nova.objects.instance [None req-6a49e111-9fe9-4008-a52a-349e6f75090d tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lazy-loading 'flavor' on Instance uuid 7183ec4a-992d-48f1-8dda-7f499c2f4e1d {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1048.088867] env[62692]: DEBUG oslo_concurrency.lockutils [None req-25e44e32-4bcf-45a3-9769-eca012e35bf1 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "d99737d8-2eb0-40ee-b61e-6c736c84ea59" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 16.635s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.151079] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "refresh_cache-a1951dba-ec5a-4849-9e22-37fba08ec60c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.151259] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquired lock "refresh_cache-a1951dba-ec5a-4849-9e22-37fba08ec60c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.151471] env[62692]: DEBUG nova.network.neutron [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1048.306077] env[62692]: DEBUG oslo_vmware.api [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141685, 'name': ReconfigVM_Task, 'duration_secs': 0.803922} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.306382] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Reconfigured VM instance instance-0000005a to attach disk [datastore2] 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da/9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1048.307012] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2479a28a-8fdb-4ac7-abc0-ecf7ad7c53b6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.313210] env[62692]: DEBUG oslo_vmware.api [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1048.313210] env[62692]: value = "task-1141686" [ 1048.313210] env[62692]: _type = "Task" [ 1048.313210] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.321703] env[62692]: DEBUG oslo_vmware.api [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141686, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.372083] env[62692]: DEBUG nova.scheduler.client.report [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1048.390507] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "d99737d8-2eb0-40ee-b61e-6c736c84ea59" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.390920] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "d99737d8-2eb0-40ee-b61e-6c736c84ea59" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.391201] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "d99737d8-2eb0-40ee-b61e-6c736c84ea59-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.391465] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "d99737d8-2eb0-40ee-b61e-6c736c84ea59-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.391688] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "d99737d8-2eb0-40ee-b61e-6c736c84ea59-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.394123] env[62692]: INFO nova.compute.manager [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Terminating instance [ 1048.395984] env[62692]: DEBUG nova.compute.manager [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1048.396205] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1048.398365] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1349a4a7-a3c5-4fd8-bca1-8f608cb8c3a4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.406274] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1048.406571] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-72d3387b-2e5f-4339-912b-b5343da39e11 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.412436] env[62692]: DEBUG oslo_vmware.api [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1048.412436] env[62692]: value = "task-1141687" [ 1048.412436] env[62692]: _type = "Task" [ 1048.412436] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.422064] env[62692]: DEBUG oslo_vmware.api [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141687, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.464720] env[62692]: DEBUG oslo_vmware.api [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]529f4465-edd2-3ee9-aaee-e8d3da948bfc, 'name': SearchDatastore_Task, 'duration_secs': 0.009499} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.465070] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.465467] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] af260f9a-4e9b-4af1-90da-c619c4755eca/af260f9a-4e9b-4af1-90da-c619c4755eca.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1048.465801] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6aba41e8-9c2d-4d77-807b-c815bc8e8773 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.471553] env[62692]: DEBUG oslo_vmware.api [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Waiting for the task: (returnval){ [ 1048.471553] env[62692]: value = "task-1141688" [ 1048.471553] env[62692]: _type = "Task" [ 1048.471553] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.479611] env[62692]: DEBUG oslo_vmware.api [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141688, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.525913] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a49e111-9fe9-4008-a52a-349e6f75090d tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1048.526244] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9857191a-7e1b-442c-a8af-b71d5af6c240 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.533828] env[62692]: DEBUG oslo_vmware.api [None req-6a49e111-9fe9-4008-a52a-349e6f75090d tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1048.533828] env[62692]: value = "task-1141689" [ 1048.533828] env[62692]: _type = "Task" [ 1048.533828] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.542409] env[62692]: DEBUG oslo_vmware.api [None req-6a49e111-9fe9-4008-a52a-349e6f75090d tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141689, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.702796] env[62692]: DEBUG nova.network.neutron [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1048.825467] env[62692]: DEBUG oslo_vmware.api [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141686, 'name': Rename_Task, 'duration_secs': 0.144675} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.825773] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1048.826042] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3c5b9602-0cbc-42c8-b3a9-ed097790ec35 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.836702] env[62692]: DEBUG oslo_vmware.api [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1048.836702] env[62692]: value = "task-1141690" [ 1048.836702] env[62692]: _type = "Task" [ 1048.836702] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.846867] env[62692]: DEBUG oslo_vmware.api [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141690, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.869188] env[62692]: DEBUG nova.network.neutron [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Updating instance_info_cache with network_info: [{"id": "5427dd91-588d-4a85-9ca4-951c8a2588be", "address": "fa:16:3e:f0:97:f0", "network": {"id": "d0e00428-1d04-4b63-8975-7f3085784358", "bridge": "br-int", "label": "tempest-ServersTestJSON-408035146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e13f8888d5e84eed8247101558e29a56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5427dd91-58", "ovs_interfaceid": "5427dd91-588d-4a85-9ca4-951c8a2588be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.877820] env[62692]: DEBUG oslo_concurrency.lockutils [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.865s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.880455] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.799s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.880790] env[62692]: DEBUG nova.objects.instance [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lazy-loading 'resources' on Instance uuid ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1048.899014] env[62692]: INFO nova.scheduler.client.report [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Deleted allocations for instance 5a08fa8a-f9fe-4879-bb7b-baa04097df6c [ 1048.928177] env[62692]: DEBUG oslo_vmware.api [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141687, 'name': PowerOffVM_Task, 'duration_secs': 0.197459} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.928539] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1048.928728] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1048.929055] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bc1d4fc8-4ebb-4b85-b8f4-8fe7ff26580c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.984295] env[62692]: DEBUG oslo_vmware.api [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141688, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.023334] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1049.023405] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1049.023575] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Deleting the datastore file [datastore2] d99737d8-2eb0-40ee-b61e-6c736c84ea59 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1049.023963] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ba161315-c284-4b77-be8e-23992d21f2e9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.031585] env[62692]: DEBUG oslo_vmware.api [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1049.031585] env[62692]: value = "task-1141692" [ 1049.031585] env[62692]: _type = "Task" [ 1049.031585] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.042783] env[62692]: DEBUG oslo_vmware.api [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141692, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.045688] env[62692]: DEBUG oslo_vmware.api [None req-6a49e111-9fe9-4008-a52a-349e6f75090d tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141689, 'name': PowerOffVM_Task, 'duration_secs': 0.174671} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.045965] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a49e111-9fe9-4008-a52a-349e6f75090d tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1049.046186] env[62692]: DEBUG nova.compute.manager [None req-6a49e111-9fe9-4008-a52a-349e6f75090d tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1049.047135] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef972f0d-0080-4c58-9c4e-f89448d18255 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.347101] env[62692]: DEBUG oslo_vmware.api [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141690, 'name': PowerOnVM_Task, 'duration_secs': 0.476662} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.347976] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1049.348572] env[62692]: INFO nova.compute.manager [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Took 7.36 seconds to spawn the instance on the hypervisor. [ 1049.348913] env[62692]: DEBUG nova.compute.manager [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1049.350989] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03fc5332-f460-42d4-bf04-9f3bac78b7d7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.374095] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Releasing lock "refresh_cache-a1951dba-ec5a-4849-9e22-37fba08ec60c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.377144] env[62692]: DEBUG nova.compute.manager [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Instance network_info: |[{"id": "5427dd91-588d-4a85-9ca4-951c8a2588be", "address": "fa:16:3e:f0:97:f0", "network": {"id": "d0e00428-1d04-4b63-8975-7f3085784358", "bridge": "br-int", "label": "tempest-ServersTestJSON-408035146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e13f8888d5e84eed8247101558e29a56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5427dd91-58", "ovs_interfaceid": "5427dd91-588d-4a85-9ca4-951c8a2588be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1049.377144] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f0:97:f0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5427dd91-588d-4a85-9ca4-951c8a2588be', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1049.390233] env[62692]: DEBUG oslo.service.loopingcall [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1049.394593] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1049.394593] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bda1e513-b07c-4c51-a29c-ff004e1c01bc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.421046] env[62692]: DEBUG oslo_concurrency.lockutils [None req-dc527d2e-5c0a-4e8b-b53d-85d034fdcfa4 tempest-ServersAdminTestJSON-661299536 tempest-ServersAdminTestJSON-661299536-project-member] Lock "5a08fa8a-f9fe-4879-bb7b-baa04097df6c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.635s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.425042] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1049.425042] env[62692]: value = "task-1141693" [ 1049.425042] env[62692]: _type = "Task" [ 1049.425042] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.438326] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141693, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.485052] env[62692]: DEBUG oslo_vmware.api [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141688, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.519832} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.488499] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] af260f9a-4e9b-4af1-90da-c619c4755eca/af260f9a-4e9b-4af1-90da-c619c4755eca.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1049.488816] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1049.489358] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-21a6132e-20e1-48ff-acbe-60d825712940 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.496472] env[62692]: DEBUG oslo_vmware.api [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Waiting for the task: (returnval){ [ 1049.496472] env[62692]: value = "task-1141694" [ 1049.496472] env[62692]: _type = "Task" [ 1049.496472] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.507580] env[62692]: DEBUG oslo_vmware.api [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141694, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.542740] env[62692]: DEBUG oslo_vmware.api [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141692, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151368} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.543035] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1049.543248] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1049.543443] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1049.543642] env[62692]: INFO nova.compute.manager [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1049.543929] env[62692]: DEBUG oslo.service.loopingcall [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1049.544158] env[62692]: DEBUG nova.compute.manager [-] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1049.544256] env[62692]: DEBUG nova.network.neutron [-] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1049.596871] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6a49e111-9fe9-4008-a52a-349e6f75090d tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "7183ec4a-992d-48f1-8dda-7f499c2f4e1d" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.549s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.649651] env[62692]: DEBUG nova.compute.manager [req-c036b110-696a-4509-9fc5-22d62f3de9ab req-4bd6e4ae-5027-4c99-808c-1a5361036f3e service nova] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Received event network-changed-5427dd91-588d-4a85-9ca4-951c8a2588be {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1049.649864] env[62692]: DEBUG nova.compute.manager [req-c036b110-696a-4509-9fc5-22d62f3de9ab req-4bd6e4ae-5027-4c99-808c-1a5361036f3e service nova] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Refreshing instance network info cache due to event network-changed-5427dd91-588d-4a85-9ca4-951c8a2588be. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1049.650413] env[62692]: DEBUG oslo_concurrency.lockutils [req-c036b110-696a-4509-9fc5-22d62f3de9ab req-4bd6e4ae-5027-4c99-808c-1a5361036f3e service nova] Acquiring lock "refresh_cache-a1951dba-ec5a-4849-9e22-37fba08ec60c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1049.650717] env[62692]: DEBUG oslo_concurrency.lockutils [req-c036b110-696a-4509-9fc5-22d62f3de9ab req-4bd6e4ae-5027-4c99-808c-1a5361036f3e service nova] Acquired lock "refresh_cache-a1951dba-ec5a-4849-9e22-37fba08ec60c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.650854] env[62692]: DEBUG nova.network.neutron [req-c036b110-696a-4509-9fc5-22d62f3de9ab req-4bd6e4ae-5027-4c99-808c-1a5361036f3e service nova] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Refreshing network info cache for port 5427dd91-588d-4a85-9ca4-951c8a2588be {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1049.681266] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b26728a5-cd0f-4767-8e9b-86a303412dbc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.694476] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-999c58a0-2196-4f7e-8825-b8b4c985911c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.739260] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f06ae0-2201-45f6-8a25-ed57cc62c09b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.751508] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e4be15e-6606-4bfe-a11e-bb81ff266f29 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.766807] env[62692]: DEBUG nova.compute.provider_tree [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1049.867699] env[62692]: INFO nova.compute.manager [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Took 25.55 seconds to build instance. [ 1049.939711] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141693, 'name': CreateVM_Task, 'duration_secs': 0.403245} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.939711] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1049.939711] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1049.939711] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.939711] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1049.939711] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae1462b7-b5eb-484c-8e6a-da42f114c673 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.943228] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 1049.943228] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5222ce4a-7b21-92a2-4937-44f13f03e9e1" [ 1049.943228] env[62692]: _type = "Task" [ 1049.943228] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.952767] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5222ce4a-7b21-92a2-4937-44f13f03e9e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.006734] env[62692]: DEBUG oslo_vmware.api [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141694, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062335} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.007068] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1050.007893] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76adfe67-6ba3-4221-942d-7ed6cf286d1f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.031362] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] af260f9a-4e9b-4af1-90da-c619c4755eca/af260f9a-4e9b-4af1-90da-c619c4755eca.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1050.032014] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-81c74552-9e5c-4e75-9cf7-69ce271cead0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.056020] env[62692]: DEBUG oslo_vmware.api [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Waiting for the task: (returnval){ [ 1050.056020] env[62692]: value = "task-1141695" [ 1050.056020] env[62692]: _type = "Task" [ 1050.056020] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.063878] env[62692]: DEBUG oslo_vmware.api [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141695, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.275324] env[62692]: DEBUG nova.scheduler.client.report [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1050.370562] env[62692]: DEBUG oslo_concurrency.lockutils [None req-94a07bc5-dce8-4783-9885-519690b5af26 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.059s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.440944] env[62692]: DEBUG nova.network.neutron [req-c036b110-696a-4509-9fc5-22d62f3de9ab req-4bd6e4ae-5027-4c99-808c-1a5361036f3e service nova] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Updated VIF entry in instance network info cache for port 5427dd91-588d-4a85-9ca4-951c8a2588be. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1050.441960] env[62692]: DEBUG nova.network.neutron [req-c036b110-696a-4509-9fc5-22d62f3de9ab req-4bd6e4ae-5027-4c99-808c-1a5361036f3e service nova] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Updating instance_info_cache with network_info: [{"id": "5427dd91-588d-4a85-9ca4-951c8a2588be", "address": "fa:16:3e:f0:97:f0", "network": {"id": "d0e00428-1d04-4b63-8975-7f3085784358", "bridge": "br-int", "label": "tempest-ServersTestJSON-408035146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e13f8888d5e84eed8247101558e29a56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5427dd91-58", "ovs_interfaceid": "5427dd91-588d-4a85-9ca4-951c8a2588be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.455696] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5222ce4a-7b21-92a2-4937-44f13f03e9e1, 'name': SearchDatastore_Task, 'duration_secs': 0.009925} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.456119] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.456286] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1050.456532] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.456684] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.457118] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1050.457784] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-354ac492-3cf4-422e-bf54-b701547ef9b8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.468048] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1050.468253] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1050.468976] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-525db054-dd74-430b-b3b4-067988da1a09 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.475271] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 1050.475271] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]528f5628-5276-552d-01f6-d3dd11357a18" [ 1050.475271] env[62692]: _type = "Task" [ 1050.475271] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.483056] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]528f5628-5276-552d-01f6-d3dd11357a18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.563921] env[62692]: DEBUG oslo_vmware.api [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141695, 'name': ReconfigVM_Task, 'duration_secs': 0.295203} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.564262] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Reconfigured VM instance instance-0000005b to attach disk [datastore2] af260f9a-4e9b-4af1-90da-c619c4755eca/af260f9a-4e9b-4af1-90da-c619c4755eca.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1050.564906] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-10e42433-1c4b-4394-a233-20916ee17c4c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.572614] env[62692]: DEBUG oslo_vmware.api [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Waiting for the task: (returnval){ [ 1050.572614] env[62692]: value = "task-1141696" [ 1050.572614] env[62692]: _type = "Task" [ 1050.572614] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.580078] env[62692]: DEBUG oslo_vmware.api [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141696, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.645964] env[62692]: DEBUG nova.network.neutron [-] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.688170] env[62692]: DEBUG nova.objects.instance [None req-9dafca49-9019-4a57-b6bc-c82e26916f39 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lazy-loading 'flavor' on Instance uuid 7183ec4a-992d-48f1-8dda-7f499c2f4e1d {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1050.780834] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.900s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.784116] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.455s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.784517] env[62692]: DEBUG nova.objects.instance [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lazy-loading 'resources' on Instance uuid 754b3c65-1e4b-49d2-8980-095d975edb01 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1050.811978] env[62692]: INFO nova.scheduler.client.report [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Deleted allocations for instance ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64 [ 1050.944674] env[62692]: DEBUG oslo_concurrency.lockutils [req-c036b110-696a-4509-9fc5-22d62f3de9ab req-4bd6e4ae-5027-4c99-808c-1a5361036f3e service nova] Releasing lock "refresh_cache-a1951dba-ec5a-4849-9e22-37fba08ec60c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.990758] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]528f5628-5276-552d-01f6-d3dd11357a18, 'name': SearchDatastore_Task, 'duration_secs': 0.008074} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.992046] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ebd068b-145d-4bcb-9be9-30dd07a180ee {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.997756] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 1050.997756] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52f6c0cc-523c-e24c-5385-3ee75a22a0dd" [ 1050.997756] env[62692]: _type = "Task" [ 1050.997756] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.008603] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52f6c0cc-523c-e24c-5385-3ee75a22a0dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.083969] env[62692]: DEBUG oslo_vmware.api [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141696, 'name': Rename_Task, 'duration_secs': 0.155636} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.084284] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1051.084548] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fada7c2e-8566-4518-95d9-bb1f2ee98069 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.092399] env[62692]: DEBUG oslo_vmware.api [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Waiting for the task: (returnval){ [ 1051.092399] env[62692]: value = "task-1141697" [ 1051.092399] env[62692]: _type = "Task" [ 1051.092399] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.101628] env[62692]: DEBUG oslo_vmware.api [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141697, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.148397] env[62692]: INFO nova.compute.manager [-] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Took 1.60 seconds to deallocate network for instance. [ 1051.194507] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9dafca49-9019-4a57-b6bc-c82e26916f39 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "refresh_cache-7183ec4a-992d-48f1-8dda-7f499c2f4e1d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.194698] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9dafca49-9019-4a57-b6bc-c82e26916f39 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquired lock "refresh_cache-7183ec4a-992d-48f1-8dda-7f499c2f4e1d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.194861] env[62692]: DEBUG nova.network.neutron [None req-9dafca49-9019-4a57-b6bc-c82e26916f39 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1051.195052] env[62692]: DEBUG nova.objects.instance [None req-9dafca49-9019-4a57-b6bc-c82e26916f39 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lazy-loading 'info_cache' on Instance uuid 7183ec4a-992d-48f1-8dda-7f499c2f4e1d {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1051.323753] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f8433cf7-3e0f-4c22-80f8-4833d552447f tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.851s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.503024] env[62692]: DEBUG nova.compute.manager [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Stashing vm_state: active {{(pid=62692) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1051.510728] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0145837c-710f-4cb1-b088-5b939278e726 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.518068] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52f6c0cc-523c-e24c-5385-3ee75a22a0dd, 'name': SearchDatastore_Task, 'duration_secs': 0.011534} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.518705] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.519058] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] a1951dba-ec5a-4849-9e22-37fba08ec60c/a1951dba-ec5a-4849-9e22-37fba08ec60c.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1051.519373] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f0ee4011-0197-49f9-82a6-181f631daeb3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.524829] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a25bba6-5d07-447e-a0af-94e4f4058581 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.530018] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 1051.530018] env[62692]: value = "task-1141698" [ 1051.530018] env[62692]: _type = "Task" [ 1051.530018] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.564306] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cf5a191-5f9b-4a1d-b53b-04a052e0120e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.567011] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141698, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.575295] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a47f93ef-d3a9-47a4-9ab6-1e55f45d713a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.591844] env[62692]: DEBUG nova.compute.provider_tree [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1051.605343] env[62692]: DEBUG oslo_vmware.api [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141697, 'name': PowerOnVM_Task, 'duration_secs': 0.4905} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.605343] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1051.605343] env[62692]: INFO nova.compute.manager [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Took 7.29 seconds to spawn the instance on the hypervisor. [ 1051.605343] env[62692]: DEBUG nova.compute.manager [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1051.606232] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03d96a10-d8fc-4e82-bbc6-28e2177f8e77 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.655643] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.693622] env[62692]: DEBUG nova.compute.manager [req-3a1750c4-ffb1-489e-af57-b693d45df5d0 req-f3d5f8bd-4353-4dbb-b3e7-48d86a1796c4 service nova] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Received event network-vif-deleted-7f2cabdf-cbda-4cf7-a518-13f619716d12 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1051.697739] env[62692]: DEBUG nova.objects.base [None req-9dafca49-9019-4a57-b6bc-c82e26916f39 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Object Instance<7183ec4a-992d-48f1-8dda-7f499c2f4e1d> lazy-loaded attributes: flavor,info_cache {{(pid=62692) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1052.044343] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141698, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.051548] env[62692]: DEBUG oslo_concurrency.lockutils [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.098664] env[62692]: DEBUG nova.scheduler.client.report [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1052.127608] env[62692]: INFO nova.compute.manager [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Took 25.75 seconds to build instance. [ 1052.479712] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "bb547773-d176-4c8e-a0fa-a374d5050b1b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.480079] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "bb547773-d176-4c8e-a0fa-a374d5050b1b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.480337] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "bb547773-d176-4c8e-a0fa-a374d5050b1b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.480527] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "bb547773-d176-4c8e-a0fa-a374d5050b1b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.480701] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "bb547773-d176-4c8e-a0fa-a374d5050b1b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.483223] env[62692]: INFO nova.compute.manager [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Terminating instance [ 1052.486338] env[62692]: DEBUG nova.compute.manager [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1052.486626] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1052.488127] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63a6e872-5856-4431-8aad-d83332113dfc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.496710] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1052.496968] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-acee72d1-2ea8-4998-95f6-83c78216dfcb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.505262] env[62692]: DEBUG oslo_vmware.api [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 1052.505262] env[62692]: value = "task-1141699" [ 1052.505262] env[62692]: _type = "Task" [ 1052.505262] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.517498] env[62692]: DEBUG oslo_vmware.api [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141699, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.545309] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141698, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.546822} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.545464] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] a1951dba-ec5a-4849-9e22-37fba08ec60c/a1951dba-ec5a-4849-9e22-37fba08ec60c.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1052.545758] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1052.546081] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-088f8e9d-bd34-410d-964f-9d86b4402335 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.554514] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 1052.554514] env[62692]: value = "task-1141700" [ 1052.554514] env[62692]: _type = "Task" [ 1052.554514] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.555555] env[62692]: DEBUG nova.network.neutron [None req-9dafca49-9019-4a57-b6bc-c82e26916f39 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Updating instance_info_cache with network_info: [{"id": "ab9d5209-bc8f-437d-a219-7f521cae10e6", "address": "fa:16:3e:d1:60:11", "network": {"id": "26fe2f12-b216-4fa5-a022-24c5ab315416", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-838584488-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb2ee16ec90c48e6babfb1d31ea3de94", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab9d5209-bc", "ovs_interfaceid": "ab9d5209-bc8f-437d-a219-7f521cae10e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1052.568998] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141700, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.605035] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.821s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.607962] env[62692]: DEBUG oslo_concurrency.lockutils [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.269s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.610642] env[62692]: INFO nova.compute.claims [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1052.629059] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0d11bffa-b324-4cf0-8edb-ff25128ceabd tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Lock "af260f9a-4e9b-4af1-90da-c619c4755eca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.266s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.632015] env[62692]: INFO nova.scheduler.client.report [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Deleted allocations for instance 754b3c65-1e4b-49d2-8980-095d975edb01 [ 1052.970752] env[62692]: DEBUG nova.compute.manager [req-1dfc3823-8b65-4b1e-ac94-642cb007a58e req-1eda1f90-a48a-43bb-ab22-8ff62bea6e5f service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Received event network-changed-c43f9978-4cfd-467d-86e0-4c8acfefc821 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1052.970968] env[62692]: DEBUG nova.compute.manager [req-1dfc3823-8b65-4b1e-ac94-642cb007a58e req-1eda1f90-a48a-43bb-ab22-8ff62bea6e5f service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Refreshing instance network info cache due to event network-changed-c43f9978-4cfd-467d-86e0-4c8acfefc821. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1052.971209] env[62692]: DEBUG oslo_concurrency.lockutils [req-1dfc3823-8b65-4b1e-ac94-642cb007a58e req-1eda1f90-a48a-43bb-ab22-8ff62bea6e5f service nova] Acquiring lock "refresh_cache-1b753882-9118-479c-aa3e-cbb8bdf2e086" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1052.971361] env[62692]: DEBUG oslo_concurrency.lockutils [req-1dfc3823-8b65-4b1e-ac94-642cb007a58e req-1eda1f90-a48a-43bb-ab22-8ff62bea6e5f service nova] Acquired lock "refresh_cache-1b753882-9118-479c-aa3e-cbb8bdf2e086" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.971550] env[62692]: DEBUG nova.network.neutron [req-1dfc3823-8b65-4b1e-ac94-642cb007a58e req-1eda1f90-a48a-43bb-ab22-8ff62bea6e5f service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Refreshing network info cache for port c43f9978-4cfd-467d-86e0-4c8acfefc821 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1053.015962] env[62692]: DEBUG oslo_vmware.api [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141699, 'name': PowerOffVM_Task, 'duration_secs': 0.367395} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.016514] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1053.016514] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1053.016657] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-80dc9979-793d-402f-b26e-d1af7a48cf95 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.065618] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9dafca49-9019-4a57-b6bc-c82e26916f39 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Releasing lock "refresh_cache-7183ec4a-992d-48f1-8dda-7f499c2f4e1d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.074788] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141700, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.222111} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.074983] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1053.076227] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83fa4f29-3be9-46ed-8933-6c9f1012c3a7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.084025] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1053.084025] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1053.084131] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Deleting the datastore file [datastore2] bb547773-d176-4c8e-a0fa-a374d5050b1b {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1053.093838] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ed9dc820-d4ba-46e8-998d-ac735322c970 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.104447] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] a1951dba-ec5a-4849-9e22-37fba08ec60c/a1951dba-ec5a-4849-9e22-37fba08ec60c.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1053.105297] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5642634f-2916-4aef-8df2-441a00c25ae8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.128674] env[62692]: DEBUG oslo_vmware.api [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for the task: (returnval){ [ 1053.128674] env[62692]: value = "task-1141702" [ 1053.128674] env[62692]: _type = "Task" [ 1053.128674] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.130626] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 1053.130626] env[62692]: value = "task-1141703" [ 1053.130626] env[62692]: _type = "Task" [ 1053.130626] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.148967] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141703, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.149537] env[62692]: DEBUG oslo_vmware.api [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141702, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.149725] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7b388f34-6580-448e-8d26-03909a6da293 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "754b3c65-1e4b-49d2-8980-095d975edb01" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.614s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.571780] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dafca49-9019-4a57-b6bc-c82e26916f39 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1053.572131] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4361ec6d-8e8e-4faa-aa97-185672f10359 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.580788] env[62692]: DEBUG oslo_vmware.api [None req-9dafca49-9019-4a57-b6bc-c82e26916f39 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1053.580788] env[62692]: value = "task-1141704" [ 1053.580788] env[62692]: _type = "Task" [ 1053.580788] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.594745] env[62692]: DEBUG oslo_vmware.api [None req-9dafca49-9019-4a57-b6bc-c82e26916f39 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141704, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.644082] env[62692]: DEBUG oslo_vmware.api [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Task: {'id': task-1141702, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157077} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.650902] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1053.651168] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1053.651382] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1053.651566] env[62692]: INFO nova.compute.manager [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1053.651831] env[62692]: DEBUG oslo.service.loopingcall [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1053.652097] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141703, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.654799] env[62692]: DEBUG nova.compute.manager [-] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1053.654901] env[62692]: DEBUG nova.network.neutron [-] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1053.734730] env[62692]: DEBUG nova.network.neutron [req-1dfc3823-8b65-4b1e-ac94-642cb007a58e req-1eda1f90-a48a-43bb-ab22-8ff62bea6e5f service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Updated VIF entry in instance network info cache for port c43f9978-4cfd-467d-86e0-4c8acfefc821. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1053.735103] env[62692]: DEBUG nova.network.neutron [req-1dfc3823-8b65-4b1e-ac94-642cb007a58e req-1eda1f90-a48a-43bb-ab22-8ff62bea6e5f service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Updating instance_info_cache with network_info: [{"id": "c43f9978-4cfd-467d-86e0-4c8acfefc821", "address": "fa:16:3e:ef:d0:33", "network": {"id": "b7c23ada-5498-4686-bd7b-b32d0837e401", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1234309205-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.238", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f26b93a4a43b45e8a11f3793996c8496", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b67e519-46cf-44ce-b670-4ba4c0c5b658", "external-id": "nsx-vlan-transportzone-110", "segmentation_id": 110, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc43f9978-4c", "ovs_interfaceid": "c43f9978-4cfd-467d-86e0-4c8acfefc821", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.867250] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92491179-89a6-4cfa-b883-9f1e29373851 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.877149] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7db5052d-a791-4700-b545-16babe051fbb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.911330] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a6bc082-04d4-456b-8744-f5ef57ab3e23 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.919655] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90d8c601-7ef5-418f-9afd-081e8d3eeb38 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.940233] env[62692]: DEBUG nova.compute.provider_tree [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1054.094553] env[62692]: DEBUG oslo_vmware.api [None req-9dafca49-9019-4a57-b6bc-c82e26916f39 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141704, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.146410] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141703, 'name': ReconfigVM_Task, 'duration_secs': 0.734122} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.146746] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Reconfigured VM instance instance-0000005c to attach disk [datastore2] a1951dba-ec5a-4849-9e22-37fba08ec60c/a1951dba-ec5a-4849-9e22-37fba08ec60c.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1054.147922] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e4bb4a44-50db-4156-8dec-4b64138f4dbe {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.156754] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 1054.156754] env[62692]: value = "task-1141705" [ 1054.156754] env[62692]: _type = "Task" [ 1054.156754] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.173810] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141705, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.238111] env[62692]: DEBUG oslo_concurrency.lockutils [req-1dfc3823-8b65-4b1e-ac94-642cb007a58e req-1eda1f90-a48a-43bb-ab22-8ff62bea6e5f service nova] Releasing lock "refresh_cache-1b753882-9118-479c-aa3e-cbb8bdf2e086" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1054.443256] env[62692]: DEBUG nova.scheduler.client.report [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1054.590661] env[62692]: DEBUG nova.network.neutron [-] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.602123] env[62692]: DEBUG oslo_vmware.api [None req-9dafca49-9019-4a57-b6bc-c82e26916f39 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141704, 'name': PowerOnVM_Task, 'duration_secs': 0.604802} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.602303] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dafca49-9019-4a57-b6bc-c82e26916f39 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1054.603114] env[62692]: DEBUG nova.compute.manager [None req-9dafca49-9019-4a57-b6bc-c82e26916f39 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1054.605401] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bd94596-6e8f-4f00-9e8f-147d32ed3f70 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.667262] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141705, 'name': Rename_Task, 'duration_secs': 0.168447} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.668094] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1054.668094] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f1b1f041-f0cc-4b71-9fcb-7722aff84ce4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.677076] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 1054.677076] env[62692]: value = "task-1141706" [ 1054.677076] env[62692]: _type = "Task" [ 1054.677076] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.688967] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141706, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.816327] env[62692]: DEBUG nova.compute.manager [req-7660724e-3a21-4447-9c7f-6d07cb9db9d6 req-047126e7-a5f1-4407-8af5-70da99fa37bf service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Received event network-changed-c43f9978-4cfd-467d-86e0-4c8acfefc821 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1054.816607] env[62692]: DEBUG nova.compute.manager [req-7660724e-3a21-4447-9c7f-6d07cb9db9d6 req-047126e7-a5f1-4407-8af5-70da99fa37bf service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Refreshing instance network info cache due to event network-changed-c43f9978-4cfd-467d-86e0-4c8acfefc821. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1054.816858] env[62692]: DEBUG oslo_concurrency.lockutils [req-7660724e-3a21-4447-9c7f-6d07cb9db9d6 req-047126e7-a5f1-4407-8af5-70da99fa37bf service nova] Acquiring lock "refresh_cache-1b753882-9118-479c-aa3e-cbb8bdf2e086" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.817116] env[62692]: DEBUG oslo_concurrency.lockutils [req-7660724e-3a21-4447-9c7f-6d07cb9db9d6 req-047126e7-a5f1-4407-8af5-70da99fa37bf service nova] Acquired lock "refresh_cache-1b753882-9118-479c-aa3e-cbb8bdf2e086" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.817461] env[62692]: DEBUG nova.network.neutron [req-7660724e-3a21-4447-9c7f-6d07cb9db9d6 req-047126e7-a5f1-4407-8af5-70da99fa37bf service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Refreshing network info cache for port c43f9978-4cfd-467d-86e0-4c8acfefc821 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1054.953431] env[62692]: DEBUG oslo_concurrency.lockutils [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.345s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.954179] env[62692]: DEBUG nova.compute.manager [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1054.956848] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.368s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.957088] env[62692]: DEBUG nova.objects.instance [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lazy-loading 'resources' on Instance uuid 668fb0ea-7eed-4198-943c-8f916ec11368 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1055.097220] env[62692]: INFO nova.compute.manager [-] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Took 1.44 seconds to deallocate network for instance. [ 1055.151645] env[62692]: DEBUG nova.compute.manager [req-52be2da0-f069-43f8-9a44-dcd153925fcc req-c59dcd76-140f-4bb9-95b1-33bb7222d7e3 service nova] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Received event network-vif-deleted-c77af962-553d-4504-825e-78a6689aee79 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1055.188343] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141706, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.459955] env[62692]: DEBUG nova.compute.utils [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1055.462177] env[62692]: DEBUG nova.compute.manager [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1055.466025] env[62692]: DEBUG nova.network.neutron [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1055.561075] env[62692]: DEBUG nova.network.neutron [req-7660724e-3a21-4447-9c7f-6d07cb9db9d6 req-047126e7-a5f1-4407-8af5-70da99fa37bf service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Updated VIF entry in instance network info cache for port c43f9978-4cfd-467d-86e0-4c8acfefc821. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1055.561075] env[62692]: DEBUG nova.network.neutron [req-7660724e-3a21-4447-9c7f-6d07cb9db9d6 req-047126e7-a5f1-4407-8af5-70da99fa37bf service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Updating instance_info_cache with network_info: [{"id": "c43f9978-4cfd-467d-86e0-4c8acfefc821", "address": "fa:16:3e:ef:d0:33", "network": {"id": "b7c23ada-5498-4686-bd7b-b32d0837e401", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1234309205-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f26b93a4a43b45e8a11f3793996c8496", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b67e519-46cf-44ce-b670-4ba4c0c5b658", "external-id": "nsx-vlan-transportzone-110", "segmentation_id": 110, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc43f9978-4c", "ovs_interfaceid": "c43f9978-4cfd-467d-86e0-4c8acfefc821", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.594073] env[62692]: DEBUG nova.policy [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a28f5e0e0af443c5be82214d446c221d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '589895ae20394e2fa525707bf7d09f2a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1055.603936] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.689813] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141706, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.710697] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a21be45e-d474-4230-b651-6ff4919b0eae {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.723723] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b58ed41d-df54-4ee2-8bfd-065b3125cc9a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.758905] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6af5f21-c496-4313-ae3e-f64b9e340dd9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.768275] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba5a825-31fd-41e6-916a-f922356be0b8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.785901] env[62692]: DEBUG nova.compute.provider_tree [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1055.965184] env[62692]: DEBUG nova.compute.manager [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1056.066706] env[62692]: DEBUG oslo_concurrency.lockutils [req-7660724e-3a21-4447-9c7f-6d07cb9db9d6 req-047126e7-a5f1-4407-8af5-70da99fa37bf service nova] Releasing lock "refresh_cache-1b753882-9118-479c-aa3e-cbb8bdf2e086" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.193366] env[62692]: DEBUG oslo_vmware.api [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141706, 'name': PowerOnVM_Task, 'duration_secs': 1.024524} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.194388] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1056.194648] env[62692]: INFO nova.compute.manager [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Took 9.56 seconds to spawn the instance on the hypervisor. [ 1056.194834] env[62692]: DEBUG nova.compute.manager [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1056.195713] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c57103a-0f6b-4791-990b-cf3bafa5d5ae {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.290794] env[62692]: DEBUG nova.scheduler.client.report [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1056.343374] env[62692]: DEBUG nova.network.neutron [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Successfully created port: 898dbc93-ff46-4a9e-9214-060a76c66507 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1056.724666] env[62692]: INFO nova.compute.manager [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Took 29.62 seconds to build instance. [ 1056.796422] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.839s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.798756] env[62692]: DEBUG oslo_concurrency.lockutils [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.747s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.802587] env[62692]: INFO nova.compute.claims [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1056.875852] env[62692]: INFO nova.scheduler.client.report [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Deleted allocations for instance 668fb0ea-7eed-4198-943c-8f916ec11368 [ 1056.977812] env[62692]: DEBUG nova.compute.manager [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1057.009753] env[62692]: DEBUG nova.virt.hardware [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1057.010055] env[62692]: DEBUG nova.virt.hardware [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1057.010285] env[62692]: DEBUG nova.virt.hardware [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1057.010532] env[62692]: DEBUG nova.virt.hardware [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1057.010717] env[62692]: DEBUG nova.virt.hardware [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1057.010915] env[62692]: DEBUG nova.virt.hardware [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1057.011245] env[62692]: DEBUG nova.virt.hardware [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1057.011448] env[62692]: DEBUG nova.virt.hardware [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1057.011604] env[62692]: DEBUG nova.virt.hardware [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1057.011777] env[62692]: DEBUG nova.virt.hardware [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1057.011956] env[62692]: DEBUG nova.virt.hardware [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1057.013239] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1533b1c8-367b-4f39-a8cf-34b8657057a2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.022020] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa288849-5ca1-496a-880f-f620118a1dad {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.228357] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ce860ed7-39c4-406b-bc76-a6596b656b4d tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "a1951dba-ec5a-4849-9e22-37fba08ec60c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.135s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.350135] env[62692]: DEBUG nova.compute.manager [req-6830d92e-2395-4339-acbe-fa879eb71ca2 req-e5a38602-b519-4532-80ca-c3da08acce74 service nova] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Received event network-changed-0b794a90-367f-4d18-a270-df14f204b2a3 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1057.350276] env[62692]: DEBUG nova.compute.manager [req-6830d92e-2395-4339-acbe-fa879eb71ca2 req-e5a38602-b519-4532-80ca-c3da08acce74 service nova] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Refreshing instance network info cache due to event network-changed-0b794a90-367f-4d18-a270-df14f204b2a3. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1057.350497] env[62692]: DEBUG oslo_concurrency.lockutils [req-6830d92e-2395-4339-acbe-fa879eb71ca2 req-e5a38602-b519-4532-80ca-c3da08acce74 service nova] Acquiring lock "refresh_cache-af260f9a-4e9b-4af1-90da-c619c4755eca" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.350655] env[62692]: DEBUG oslo_concurrency.lockutils [req-6830d92e-2395-4339-acbe-fa879eb71ca2 req-e5a38602-b519-4532-80ca-c3da08acce74 service nova] Acquired lock "refresh_cache-af260f9a-4e9b-4af1-90da-c619c4755eca" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.350815] env[62692]: DEBUG nova.network.neutron [req-6830d92e-2395-4339-acbe-fa879eb71ca2 req-e5a38602-b519-4532-80ca-c3da08acce74 service nova] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Refreshing network info cache for port 0b794a90-367f-4d18-a270-df14f204b2a3 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1057.388626] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9c708a75-def9-4893-a7f0-55c98263683c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "668fb0ea-7eed-4198-943c-8f916ec11368" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.386s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.389814] env[62692]: DEBUG oslo_concurrency.lockutils [req-5cee4abe-6825-4192-87eb-840645895468 req-cc820281-ec83-4089-8240-6caf6fa82aba service nova] Acquired lock "668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.391850] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82e3437a-fe44-43b1-95e6-95f41675ec47 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.403768] env[62692]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 1057.404086] env[62692]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=62692) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 1057.404652] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-542550fc-69e5-41d0-b25b-0cfebec00649 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.417902] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17ca7d85-250e-4ee0-bc61-d3029fda31e5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.435453] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.437762] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.462943] env[62692]: ERROR root [req-5cee4abe-6825-4192-87eb-840645895468 req-cc820281-ec83-4089-8240-6caf6fa82aba service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-249003' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-249003' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-249003' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-249003'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-249003' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-249003' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-249003'}\n"]: nova.exception.InstanceNotFound: Instance 668fb0ea-7eed-4198-943c-8f916ec11368 could not be found. [ 1057.463173] env[62692]: DEBUG oslo_concurrency.lockutils [req-5cee4abe-6825-4192-87eb-840645895468 req-cc820281-ec83-4089-8240-6caf6fa82aba service nova] Releasing lock "668fb0ea-7eed-4198-943c-8f916ec11368" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.463386] env[62692]: DEBUG nova.compute.manager [req-5cee4abe-6825-4192-87eb-840645895468 req-cc820281-ec83-4089-8240-6caf6fa82aba service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Detach interface failed, port_id=e1a45cde-8d94-464d-8c3e-e3f94b62f9e6, reason: Instance 668fb0ea-7eed-4198-943c-8f916ec11368 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1057.463635] env[62692]: DEBUG nova.compute.manager [req-5cee4abe-6825-4192-87eb-840645895468 req-cc820281-ec83-4089-8240-6caf6fa82aba service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Received event network-vif-deleted-dac9e9e1-d8ad-4ee7-8836-a84a7cb3cf27 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1057.463848] env[62692]: INFO nova.compute.manager [req-5cee4abe-6825-4192-87eb-840645895468 req-cc820281-ec83-4089-8240-6caf6fa82aba service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Neutron deleted interface dac9e9e1-d8ad-4ee7-8836-a84a7cb3cf27; detaching it from the instance and deleting it from the info cache [ 1057.464136] env[62692]: DEBUG nova.network.neutron [req-5cee4abe-6825-4192-87eb-840645895468 req-cc820281-ec83-4089-8240-6caf6fa82aba service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Updating instance_info_cache with network_info: [{"id": "2deb4259-2f64-4f7c-9289-eba76375335a", "address": "fa:16:3e:bd:3f:b3", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.135", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2deb4259-2f", "ovs_interfaceid": "2deb4259-2f64-4f7c-9289-eba76375335a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.944684] env[62692]: DEBUG nova.compute.manager [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1057.972583] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-40f066a5-d032-4801-bf07-2c5e6f6d081b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.991867] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8141f69d-b0ab-4438-bf6d-e9fd1e5b08d8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.034949] env[62692]: DEBUG nova.compute.manager [req-5cee4abe-6825-4192-87eb-840645895468 req-cc820281-ec83-4089-8240-6caf6fa82aba service nova] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Detach interface failed, port_id=dac9e9e1-d8ad-4ee7-8836-a84a7cb3cf27, reason: Instance 668fb0ea-7eed-4198-943c-8f916ec11368 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1058.035244] env[62692]: DEBUG nova.compute.manager [req-5cee4abe-6825-4192-87eb-840645895468 req-cc820281-ec83-4089-8240-6caf6fa82aba service nova] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Received event network-changed-877ab637-ea08-499f-9d34-88ca15e1ed1a {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1058.035423] env[62692]: DEBUG nova.compute.manager [req-5cee4abe-6825-4192-87eb-840645895468 req-cc820281-ec83-4089-8240-6caf6fa82aba service nova] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Refreshing instance network info cache due to event network-changed-877ab637-ea08-499f-9d34-88ca15e1ed1a. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1058.035642] env[62692]: DEBUG oslo_concurrency.lockutils [req-5cee4abe-6825-4192-87eb-840645895468 req-cc820281-ec83-4089-8240-6caf6fa82aba service nova] Acquiring lock "refresh_cache-e8f645da-3b4a-4f21-9a17-aebdc8d112df" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.035846] env[62692]: DEBUG oslo_concurrency.lockutils [req-5cee4abe-6825-4192-87eb-840645895468 req-cc820281-ec83-4089-8240-6caf6fa82aba service nova] Acquired lock "refresh_cache-e8f645da-3b4a-4f21-9a17-aebdc8d112df" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.036051] env[62692]: DEBUG nova.network.neutron [req-5cee4abe-6825-4192-87eb-840645895468 req-cc820281-ec83-4089-8240-6caf6fa82aba service nova] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Refreshing network info cache for port 877ab637-ea08-499f-9d34-88ca15e1ed1a {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1058.100267] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c13683e-5b94-40df-85b5-2ea364a32f2a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.108985] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13f52fbe-9b6e-422d-a90e-0546d840e5ee {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.149448] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dffb889b-359b-47ea-af58-b50a8e3011ac {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.161871] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0778f70-a233-4c03-8fa1-89fee664ec6d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.181713] env[62692]: DEBUG nova.compute.provider_tree [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1058.219810] env[62692]: DEBUG nova.network.neutron [req-6830d92e-2395-4339-acbe-fa879eb71ca2 req-e5a38602-b519-4532-80ca-c3da08acce74 service nova] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Updated VIF entry in instance network info cache for port 0b794a90-367f-4d18-a270-df14f204b2a3. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1058.220216] env[62692]: DEBUG nova.network.neutron [req-6830d92e-2395-4339-acbe-fa879eb71ca2 req-e5a38602-b519-4532-80ca-c3da08acce74 service nova] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Updating instance_info_cache with network_info: [{"id": "0b794a90-367f-4d18-a270-df14f204b2a3", "address": "fa:16:3e:45:cd:d0", "network": {"id": "b7c23ada-5498-4686-bd7b-b32d0837e401", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1234309205-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f26b93a4a43b45e8a11f3793996c8496", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b67e519-46cf-44ce-b670-4ba4c0c5b658", "external-id": "nsx-vlan-transportzone-110", "segmentation_id": 110, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b794a90-36", "ovs_interfaceid": "0b794a90-367f-4d18-a270-df14f204b2a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.228549] env[62692]: DEBUG oslo_concurrency.lockutils [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquiring lock "af260f9a-4e9b-4af1-90da-c619c4755eca" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.228824] env[62692]: DEBUG oslo_concurrency.lockutils [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Lock "af260f9a-4e9b-4af1-90da-c619c4755eca" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.229062] env[62692]: DEBUG oslo_concurrency.lockutils [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquiring lock "af260f9a-4e9b-4af1-90da-c619c4755eca-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.229261] env[62692]: DEBUG oslo_concurrency.lockutils [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Lock "af260f9a-4e9b-4af1-90da-c619c4755eca-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.230166] env[62692]: DEBUG oslo_concurrency.lockutils [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Lock "af260f9a-4e9b-4af1-90da-c619c4755eca-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.232467] env[62692]: INFO nova.compute.manager [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Terminating instance [ 1058.236936] env[62692]: DEBUG nova.compute.manager [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1058.237155] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1058.238026] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52e8857d-3428-4819-be98-56738416a5ba {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.248195] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1058.248483] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-32cfaf50-f736-4cc9-b9ee-d9391cc2cf68 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.256540] env[62692]: DEBUG oslo_vmware.api [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Waiting for the task: (returnval){ [ 1058.256540] env[62692]: value = "task-1141707" [ 1058.256540] env[62692]: _type = "Task" [ 1058.256540] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.268327] env[62692]: DEBUG oslo_vmware.api [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141707, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.403209] env[62692]: DEBUG nova.network.neutron [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Successfully updated port: 898dbc93-ff46-4a9e-9214-060a76c66507 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1058.475390] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.684179] env[62692]: DEBUG nova.scheduler.client.report [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1058.723019] env[62692]: DEBUG oslo_concurrency.lockutils [req-6830d92e-2395-4339-acbe-fa879eb71ca2 req-e5a38602-b519-4532-80ca-c3da08acce74 service nova] Releasing lock "refresh_cache-af260f9a-4e9b-4af1-90da-c619c4755eca" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.770075] env[62692]: DEBUG oslo_vmware.api [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141707, 'name': PowerOffVM_Task, 'duration_secs': 0.309157} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.770954] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1058.771252] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1058.771437] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d3016bfd-f612-4c42-bf28-c6667b9078a4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.853733] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1058.853950] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1058.854155] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Deleting the datastore file [datastore2] af260f9a-4e9b-4af1-90da-c619c4755eca {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1058.854473] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0690ea18-aee1-48c3-b3fe-a287a51f5a14 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.863783] env[62692]: DEBUG oslo_vmware.api [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Waiting for the task: (returnval){ [ 1058.863783] env[62692]: value = "task-1141709" [ 1058.863783] env[62692]: _type = "Task" [ 1058.863783] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.876133] env[62692]: DEBUG oslo_vmware.api [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141709, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.909978] env[62692]: DEBUG oslo_concurrency.lockutils [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "refresh_cache-ca56c2ec-95fc-44fc-856f-e395bda1703c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.909978] env[62692]: DEBUG oslo_concurrency.lockutils [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired lock "refresh_cache-ca56c2ec-95fc-44fc-856f-e395bda1703c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.909978] env[62692]: DEBUG nova.network.neutron [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1059.127757] env[62692]: DEBUG nova.network.neutron [req-5cee4abe-6825-4192-87eb-840645895468 req-cc820281-ec83-4089-8240-6caf6fa82aba service nova] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Updated VIF entry in instance network info cache for port 877ab637-ea08-499f-9d34-88ca15e1ed1a. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1059.129021] env[62692]: DEBUG nova.network.neutron [req-5cee4abe-6825-4192-87eb-840645895468 req-cc820281-ec83-4089-8240-6caf6fa82aba service nova] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Updating instance_info_cache with network_info: [{"id": "877ab637-ea08-499f-9d34-88ca15e1ed1a", "address": "fa:16:3e:73:e7:26", "network": {"id": "2febb660-329d-403c-8c2c-f11c0f5f547c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1806650357-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.233", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37921f67664544478cb2568def068b11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap877ab637-ea", "ovs_interfaceid": "877ab637-ea08-499f-9d34-88ca15e1ed1a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.192713] env[62692]: DEBUG oslo_concurrency.lockutils [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.394s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.193281] env[62692]: DEBUG nova.compute.manager [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1059.195988] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.953s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.197907] env[62692]: INFO nova.compute.claims [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1059.376911] env[62692]: DEBUG oslo_vmware.api [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141709, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154364} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.377651] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1059.379176] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1059.381284] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1059.381284] env[62692]: INFO nova.compute.manager [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1059.381284] env[62692]: DEBUG oslo.service.loopingcall [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1059.381284] env[62692]: DEBUG nova.compute.manager [-] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1059.381905] env[62692]: DEBUG nova.network.neutron [-] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1059.406494] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "a1951dba-ec5a-4849-9e22-37fba08ec60c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.406941] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "a1951dba-ec5a-4849-9e22-37fba08ec60c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.407037] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "a1951dba-ec5a-4849-9e22-37fba08ec60c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.407217] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "a1951dba-ec5a-4849-9e22-37fba08ec60c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.407385] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "a1951dba-ec5a-4849-9e22-37fba08ec60c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.411890] env[62692]: INFO nova.compute.manager [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Terminating instance [ 1059.414440] env[62692]: DEBUG nova.compute.manager [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1059.414723] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1059.415577] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b51dd01-ad39-45e6-91ca-2d9c88c19da5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.425051] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1059.425658] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ac289e8f-4438-485b-baba-123a5ac38577 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.436518] env[62692]: DEBUG oslo_vmware.api [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 1059.436518] env[62692]: value = "task-1141710" [ 1059.436518] env[62692]: _type = "Task" [ 1059.436518] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.446781] env[62692]: DEBUG oslo_vmware.api [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141710, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.470566] env[62692]: DEBUG nova.network.neutron [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1059.632739] env[62692]: DEBUG oslo_concurrency.lockutils [req-5cee4abe-6825-4192-87eb-840645895468 req-cc820281-ec83-4089-8240-6caf6fa82aba service nova] Releasing lock "refresh_cache-e8f645da-3b4a-4f21-9a17-aebdc8d112df" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.702828] env[62692]: DEBUG nova.compute.utils [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1059.704364] env[62692]: DEBUG nova.compute.manager [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1059.704513] env[62692]: DEBUG nova.network.neutron [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1059.786609] env[62692]: DEBUG nova.policy [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6102acb4c19642d8a8190d828f5b5dfc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '697b4281872f4547ad3d051e644b95ce', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1059.797021] env[62692]: DEBUG nova.compute.manager [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Received event network-changed-0b794a90-367f-4d18-a270-df14f204b2a3 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1059.797021] env[62692]: DEBUG nova.compute.manager [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Refreshing instance network info cache due to event network-changed-0b794a90-367f-4d18-a270-df14f204b2a3. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1059.797021] env[62692]: DEBUG oslo_concurrency.lockutils [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] Acquiring lock "refresh_cache-af260f9a-4e9b-4af1-90da-c619c4755eca" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.797021] env[62692]: DEBUG oslo_concurrency.lockutils [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] Acquired lock "refresh_cache-af260f9a-4e9b-4af1-90da-c619c4755eca" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.797021] env[62692]: DEBUG nova.network.neutron [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Refreshing network info cache for port 0b794a90-367f-4d18-a270-df14f204b2a3 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1059.800307] env[62692]: DEBUG nova.network.neutron [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Updating instance_info_cache with network_info: [{"id": "898dbc93-ff46-4a9e-9214-060a76c66507", "address": "fa:16:3e:bf:8f:f5", "network": {"id": "dd260020-2d74-46ba-aa78-af29bdefa176", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-617179655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "589895ae20394e2fa525707bf7d09f2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap898dbc93-ff", "ovs_interfaceid": "898dbc93-ff46-4a9e-9214-060a76c66507", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.952072] env[62692]: DEBUG oslo_vmware.api [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141710, 'name': PowerOffVM_Task, 'duration_secs': 0.241771} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.952534] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1059.952862] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1059.953237] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b405b778-ed84-4169-b3d4-ebfcf1decd7b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.040023] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1060.040023] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1060.040023] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Deleting the datastore file [datastore2] a1951dba-ec5a-4849-9e22-37fba08ec60c {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1060.040023] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-acc06267-10f5-48e7-91fe-f156d13df139 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.046192] env[62692]: DEBUG oslo_vmware.api [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 1060.046192] env[62692]: value = "task-1141712" [ 1060.046192] env[62692]: _type = "Task" [ 1060.046192] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.056955] env[62692]: DEBUG oslo_vmware.api [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141712, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.210667] env[62692]: DEBUG nova.compute.manager [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1060.219975] env[62692]: DEBUG nova.network.neutron [-] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.298458] env[62692]: DEBUG nova.network.neutron [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Successfully created port: 21e76034-6c0f-4938-afda-0d005cf45759 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1060.304264] env[62692]: DEBUG oslo_concurrency.lockutils [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Releasing lock "refresh_cache-ca56c2ec-95fc-44fc-856f-e395bda1703c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.304567] env[62692]: DEBUG nova.compute.manager [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Instance network_info: |[{"id": "898dbc93-ff46-4a9e-9214-060a76c66507", "address": "fa:16:3e:bf:8f:f5", "network": {"id": "dd260020-2d74-46ba-aa78-af29bdefa176", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-617179655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "589895ae20394e2fa525707bf7d09f2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap898dbc93-ff", "ovs_interfaceid": "898dbc93-ff46-4a9e-9214-060a76c66507", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1060.305355] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bf:8f:f5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55764410-260e-4339-a020-6b30995584bf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '898dbc93-ff46-4a9e-9214-060a76c66507', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1060.319563] env[62692]: DEBUG oslo.service.loopingcall [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1060.323619] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1060.325015] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c98fa81c-47f6-4904-a87b-143222383709 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.342912] env[62692]: INFO nova.network.neutron [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Port 0b794a90-367f-4d18-a270-df14f204b2a3 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1060.343169] env[62692]: DEBUG nova.network.neutron [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.351671] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1060.351671] env[62692]: value = "task-1141713" [ 1060.351671] env[62692]: _type = "Task" [ 1060.351671] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.362229] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141713, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.473445] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6183a6cd-e91a-43e3-af25-fa56f623b710 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.484418] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-622f4b05-a5ed-498f-b26f-1e6ef119e796 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.528850] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a37d8ab8-400e-404c-80d2-35dea193a847 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.542410] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34773ecc-efd5-44e2-8f82-ce5598e8ee0e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.559790] env[62692]: DEBUG nova.compute.provider_tree [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1060.567174] env[62692]: DEBUG oslo_vmware.api [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141712, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.337815} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.567509] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1060.567754] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1060.567995] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1060.568575] env[62692]: INFO nova.compute.manager [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1060.568575] env[62692]: DEBUG oslo.service.loopingcall [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1060.568747] env[62692]: DEBUG nova.compute.manager [-] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1060.568850] env[62692]: DEBUG nova.network.neutron [-] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1060.724218] env[62692]: INFO nova.compute.manager [-] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Took 1.34 seconds to deallocate network for instance. [ 1060.846749] env[62692]: DEBUG oslo_concurrency.lockutils [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] Releasing lock "refresh_cache-af260f9a-4e9b-4af1-90da-c619c4755eca" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.847116] env[62692]: DEBUG nova.compute.manager [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Received event network-vif-plugged-898dbc93-ff46-4a9e-9214-060a76c66507 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1060.847354] env[62692]: DEBUG oslo_concurrency.lockutils [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] Acquiring lock "ca56c2ec-95fc-44fc-856f-e395bda1703c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.847643] env[62692]: DEBUG oslo_concurrency.lockutils [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] Lock "ca56c2ec-95fc-44fc-856f-e395bda1703c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.847858] env[62692]: DEBUG oslo_concurrency.lockutils [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] Lock "ca56c2ec-95fc-44fc-856f-e395bda1703c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.848171] env[62692]: DEBUG nova.compute.manager [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] No waiting events found dispatching network-vif-plugged-898dbc93-ff46-4a9e-9214-060a76c66507 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1060.848370] env[62692]: WARNING nova.compute.manager [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Received unexpected event network-vif-plugged-898dbc93-ff46-4a9e-9214-060a76c66507 for instance with vm_state building and task_state spawning. [ 1060.848644] env[62692]: DEBUG nova.compute.manager [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Received event network-changed-898dbc93-ff46-4a9e-9214-060a76c66507 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1060.848712] env[62692]: DEBUG nova.compute.manager [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Refreshing instance network info cache due to event network-changed-898dbc93-ff46-4a9e-9214-060a76c66507. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1060.848904] env[62692]: DEBUG oslo_concurrency.lockutils [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] Acquiring lock "refresh_cache-ca56c2ec-95fc-44fc-856f-e395bda1703c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.849124] env[62692]: DEBUG oslo_concurrency.lockutils [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] Acquired lock "refresh_cache-ca56c2ec-95fc-44fc-856f-e395bda1703c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.849427] env[62692]: DEBUG nova.network.neutron [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Refreshing network info cache for port 898dbc93-ff46-4a9e-9214-060a76c66507 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1060.862391] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141713, 'name': CreateVM_Task, 'duration_secs': 0.369172} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.863357] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1060.864217] env[62692]: DEBUG oslo_concurrency.lockutils [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.864331] env[62692]: DEBUG oslo_concurrency.lockutils [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.864604] env[62692]: DEBUG oslo_concurrency.lockutils [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1060.867833] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8f8aa7b-7fb7-4f7c-8eac-0f795deb94f2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.870915] env[62692]: DEBUG oslo_vmware.api [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1060.870915] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52f36ffa-074a-d8dd-3850-b398dc7c1d39" [ 1060.870915] env[62692]: _type = "Task" [ 1060.870915] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.882556] env[62692]: DEBUG oslo_vmware.api [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52f36ffa-074a-d8dd-3850-b398dc7c1d39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.063055] env[62692]: DEBUG nova.scheduler.client.report [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1061.221403] env[62692]: DEBUG nova.compute.manager [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1061.230885] env[62692]: DEBUG oslo_concurrency.lockutils [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.244623] env[62692]: DEBUG nova.virt.hardware [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1061.244820] env[62692]: DEBUG nova.virt.hardware [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1061.245051] env[62692]: DEBUG nova.virt.hardware [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1061.245313] env[62692]: DEBUG nova.virt.hardware [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1061.245508] env[62692]: DEBUG nova.virt.hardware [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1061.246208] env[62692]: DEBUG nova.virt.hardware [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1061.246208] env[62692]: DEBUG nova.virt.hardware [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1061.246395] env[62692]: DEBUG nova.virt.hardware [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1061.246539] env[62692]: DEBUG nova.virt.hardware [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1061.246742] env[62692]: DEBUG nova.virt.hardware [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1061.246955] env[62692]: DEBUG nova.virt.hardware [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1061.247842] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfcac18e-96b6-4093-9bd0-83db4c54a67d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.252940] env[62692]: DEBUG nova.compute.manager [req-0e8372c8-5727-43ad-a350-c68117034db3 req-c52ec854-7006-46aa-bf96-1cbbe6b6a21a service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Received event network-changed-c43f9978-4cfd-467d-86e0-4c8acfefc821 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1061.253155] env[62692]: DEBUG nova.compute.manager [req-0e8372c8-5727-43ad-a350-c68117034db3 req-c52ec854-7006-46aa-bf96-1cbbe6b6a21a service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Refreshing instance network info cache due to event network-changed-c43f9978-4cfd-467d-86e0-4c8acfefc821. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1061.253368] env[62692]: DEBUG oslo_concurrency.lockutils [req-0e8372c8-5727-43ad-a350-c68117034db3 req-c52ec854-7006-46aa-bf96-1cbbe6b6a21a service nova] Acquiring lock "refresh_cache-1b753882-9118-479c-aa3e-cbb8bdf2e086" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.253493] env[62692]: DEBUG oslo_concurrency.lockutils [req-0e8372c8-5727-43ad-a350-c68117034db3 req-c52ec854-7006-46aa-bf96-1cbbe6b6a21a service nova] Acquired lock "refresh_cache-1b753882-9118-479c-aa3e-cbb8bdf2e086" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.253658] env[62692]: DEBUG nova.network.neutron [req-0e8372c8-5727-43ad-a350-c68117034db3 req-c52ec854-7006-46aa-bf96-1cbbe6b6a21a service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Refreshing network info cache for port c43f9978-4cfd-467d-86e0-4c8acfefc821 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1061.261501] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a0ea94b-7eb4-407a-9fa0-464bf98d2198 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.384499] env[62692]: DEBUG oslo_vmware.api [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52f36ffa-074a-d8dd-3850-b398dc7c1d39, 'name': SearchDatastore_Task, 'duration_secs': 0.010671} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.385238] env[62692]: DEBUG oslo_concurrency.lockutils [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.385238] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1061.385238] env[62692]: DEBUG oslo_concurrency.lockutils [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.385424] env[62692]: DEBUG oslo_concurrency.lockutils [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.385522] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1061.385794] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e30423f1-3ba1-4314-b81e-220be5734b5d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.395944] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1061.395944] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1061.396502] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-607403f6-3e6c-4c83-8ed4-90895f35554c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.403825] env[62692]: DEBUG oslo_vmware.api [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1061.403825] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52f6c403-1470-d278-473d-fa423d812b84" [ 1061.403825] env[62692]: _type = "Task" [ 1061.403825] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.416253] env[62692]: DEBUG oslo_vmware.api [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52f6c403-1470-d278-473d-fa423d812b84, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.424196] env[62692]: DEBUG nova.network.neutron [-] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.571152] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.375s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.571753] env[62692]: DEBUG nova.compute.manager [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1061.574305] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.919s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.574506] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.580039] env[62692]: DEBUG oslo_concurrency.lockutils [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 9.525s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.602021] env[62692]: INFO nova.scheduler.client.report [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Deleted allocations for instance d99737d8-2eb0-40ee-b61e-6c736c84ea59 [ 1061.722837] env[62692]: DEBUG nova.network.neutron [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Updated VIF entry in instance network info cache for port 898dbc93-ff46-4a9e-9214-060a76c66507. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1061.723832] env[62692]: DEBUG nova.network.neutron [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Updating instance_info_cache with network_info: [{"id": "898dbc93-ff46-4a9e-9214-060a76c66507", "address": "fa:16:3e:bf:8f:f5", "network": {"id": "dd260020-2d74-46ba-aa78-af29bdefa176", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-617179655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "589895ae20394e2fa525707bf7d09f2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap898dbc93-ff", "ovs_interfaceid": "898dbc93-ff46-4a9e-9214-060a76c66507", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.831153] env[62692]: DEBUG nova.compute.manager [req-ec1a3313-ca73-4d95-b589-c122bdb44017 req-4c48f988-c20d-4ea4-9721-1a05b33e4343 service nova] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Received event network-vif-deleted-0b794a90-367f-4d18-a270-df14f204b2a3 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1061.842838] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2650f739-3dd3-49fb-ae92-85115406f198 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquiring lock "af260f9a-4e9b-4af1-90da-c619c4755eca" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.895903] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "94a07505-661f-4595-9024-8a9da9f2c2fc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.896154] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "94a07505-661f-4595-9024-8a9da9f2c2fc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.917411] env[62692]: DEBUG oslo_vmware.api [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52f6c403-1470-d278-473d-fa423d812b84, 'name': SearchDatastore_Task, 'duration_secs': 0.011352} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.918230] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aacc5952-1682-4140-85bb-854cf26617a4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.924263] env[62692]: DEBUG oslo_vmware.api [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1061.924263] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52e083c5-03d6-45b2-1f28-2f6cde06e583" [ 1061.924263] env[62692]: _type = "Task" [ 1061.924263] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.928552] env[62692]: INFO nova.compute.manager [-] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Took 1.36 seconds to deallocate network for instance. [ 1061.939685] env[62692]: DEBUG oslo_vmware.api [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52e083c5-03d6-45b2-1f28-2f6cde06e583, 'name': SearchDatastore_Task, 'duration_secs': 0.01036} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.940195] env[62692]: DEBUG oslo_concurrency.lockutils [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.940455] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] ca56c2ec-95fc-44fc-856f-e395bda1703c/ca56c2ec-95fc-44fc-856f-e395bda1703c.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1061.940715] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2a08ce91-eb33-4eec-9a58-609abd84654c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.949968] env[62692]: DEBUG oslo_vmware.api [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1061.949968] env[62692]: value = "task-1141714" [ 1061.949968] env[62692]: _type = "Task" [ 1061.949968] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.960198] env[62692]: DEBUG oslo_vmware.api [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141714, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.063665] env[62692]: DEBUG nova.network.neutron [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Successfully updated port: 21e76034-6c0f-4938-afda-0d005cf45759 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1062.080782] env[62692]: DEBUG nova.compute.utils [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1062.085503] env[62692]: INFO nova.compute.claims [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1062.089832] env[62692]: DEBUG nova.compute.manager [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Not allocating networking since 'none' was specified. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 1062.108454] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d4398c09-b23a-41f4-98c8-98cd5b6d7736 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "d99737d8-2eb0-40ee-b61e-6c736c84ea59" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.717s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.115081] env[62692]: DEBUG nova.network.neutron [req-0e8372c8-5727-43ad-a350-c68117034db3 req-c52ec854-7006-46aa-bf96-1cbbe6b6a21a service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Updated VIF entry in instance network info cache for port c43f9978-4cfd-467d-86e0-4c8acfefc821. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1062.115456] env[62692]: DEBUG nova.network.neutron [req-0e8372c8-5727-43ad-a350-c68117034db3 req-c52ec854-7006-46aa-bf96-1cbbe6b6a21a service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Updating instance_info_cache with network_info: [{"id": "c43f9978-4cfd-467d-86e0-4c8acfefc821", "address": "fa:16:3e:ef:d0:33", "network": {"id": "b7c23ada-5498-4686-bd7b-b32d0837e401", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1234309205-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f26b93a4a43b45e8a11f3793996c8496", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b67e519-46cf-44ce-b670-4ba4c0c5b658", "external-id": "nsx-vlan-transportzone-110", "segmentation_id": 110, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc43f9978-4c", "ovs_interfaceid": "c43f9978-4cfd-467d-86e0-4c8acfefc821", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.226290] env[62692]: DEBUG oslo_concurrency.lockutils [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] Releasing lock "refresh_cache-ca56c2ec-95fc-44fc-856f-e395bda1703c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.226630] env[62692]: DEBUG nova.compute.manager [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Received event network-changed-c43f9978-4cfd-467d-86e0-4c8acfefc821 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1062.226757] env[62692]: DEBUG nova.compute.manager [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Refreshing instance network info cache due to event network-changed-c43f9978-4cfd-467d-86e0-4c8acfefc821. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1062.226957] env[62692]: DEBUG oslo_concurrency.lockutils [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] Acquiring lock "refresh_cache-1b753882-9118-479c-aa3e-cbb8bdf2e086" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.399923] env[62692]: DEBUG nova.compute.manager [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1062.441747] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.461288] env[62692]: DEBUG oslo_vmware.api [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141714, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.487623} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.461578] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] ca56c2ec-95fc-44fc-856f-e395bda1703c/ca56c2ec-95fc-44fc-856f-e395bda1703c.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1062.461800] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1062.462064] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-20df012b-92b6-4fa8-8cdf-b7d711d34d84 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.469574] env[62692]: DEBUG oslo_vmware.api [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1062.469574] env[62692]: value = "task-1141715" [ 1062.469574] env[62692]: _type = "Task" [ 1062.469574] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.479376] env[62692]: DEBUG oslo_vmware.api [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141715, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.567359] env[62692]: DEBUG oslo_concurrency.lockutils [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "refresh_cache-90932f05-fad3-4c6d-87ae-ab059351b0be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.567522] env[62692]: DEBUG oslo_concurrency.lockutils [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquired lock "refresh_cache-90932f05-fad3-4c6d-87ae-ab059351b0be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.567695] env[62692]: DEBUG nova.network.neutron [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1062.590632] env[62692]: DEBUG nova.compute.manager [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1062.595160] env[62692]: INFO nova.compute.resource_tracker [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Updating resource usage from migration cfaf78e4-d48c-45bb-be40-e8e1c88b83f0 [ 1062.619030] env[62692]: DEBUG oslo_concurrency.lockutils [req-0e8372c8-5727-43ad-a350-c68117034db3 req-c52ec854-7006-46aa-bf96-1cbbe6b6a21a service nova] Releasing lock "refresh_cache-1b753882-9118-479c-aa3e-cbb8bdf2e086" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.624126] env[62692]: DEBUG oslo_concurrency.lockutils [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] Acquired lock "refresh_cache-1b753882-9118-479c-aa3e-cbb8bdf2e086" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.624126] env[62692]: DEBUG nova.network.neutron [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Refreshing network info cache for port c43f9978-4cfd-467d-86e0-4c8acfefc821 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1062.783726] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4be5052d-40af-4240-9168-7e7204865f08 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.792666] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23be5bb1-253b-4394-9e82-c696253bce7f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.823820] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d9cd8bf-d377-4c3c-9f7a-f5fc79ac260e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.832550] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afc779ff-48d9-42a7-9ddf-e412f9938c4b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.847026] env[62692]: DEBUG nova.compute.provider_tree [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1062.919812] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.979696] env[62692]: DEBUG oslo_vmware.api [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141715, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064766} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.979980] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1062.980807] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b159cd11-2880-465e-9dcd-93500d93a31b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.006432] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] ca56c2ec-95fc-44fc-856f-e395bda1703c/ca56c2ec-95fc-44fc-856f-e395bda1703c.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1063.006866] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a954e56f-6159-4be4-9359-29023fb4d9d0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.032190] env[62692]: DEBUG oslo_vmware.api [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1063.032190] env[62692]: value = "task-1141716" [ 1063.032190] env[62692]: _type = "Task" [ 1063.032190] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.042399] env[62692]: DEBUG oslo_vmware.api [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141716, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.110105] env[62692]: DEBUG nova.network.neutron [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1063.272917] env[62692]: DEBUG nova.compute.manager [req-8f1cf7a0-28cf-46a9-9c63-49c980afecb5 req-a8e07239-42d3-4f27-ac6c-b5fb2a3dbd94 service nova] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Received event network-vif-deleted-5427dd91-588d-4a85-9ca4-951c8a2588be {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1063.298696] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.298696] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.349753] env[62692]: DEBUG nova.scheduler.client.report [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1063.372650] env[62692]: DEBUG nova.network.neutron [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Updating instance_info_cache with network_info: [{"id": "21e76034-6c0f-4938-afda-0d005cf45759", "address": "fa:16:3e:2d:c0:4a", "network": {"id": "77695f27-3027-4a92-8b9b-0a90cdc53f62", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-337681783-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697b4281872f4547ad3d051e644b95ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21e76034-6c", "ovs_interfaceid": "21e76034-6c0f-4938-afda-0d005cf45759", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.374717] env[62692]: DEBUG nova.network.neutron [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Updated VIF entry in instance network info cache for port c43f9978-4cfd-467d-86e0-4c8acfefc821. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1063.375106] env[62692]: DEBUG nova.network.neutron [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Updating instance_info_cache with network_info: [{"id": "c43f9978-4cfd-467d-86e0-4c8acfefc821", "address": "fa:16:3e:ef:d0:33", "network": {"id": "b7c23ada-5498-4686-bd7b-b32d0837e401", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1234309205-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f26b93a4a43b45e8a11f3793996c8496", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b67e519-46cf-44ce-b670-4ba4c0c5b658", "external-id": "nsx-vlan-transportzone-110", "segmentation_id": 110, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc43f9978-4c", "ovs_interfaceid": "c43f9978-4cfd-467d-86e0-4c8acfefc821", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.543597] env[62692]: DEBUG oslo_vmware.api [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141716, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.604947] env[62692]: DEBUG nova.compute.manager [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1063.629048] env[62692]: DEBUG nova.virt.hardware [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1063.629325] env[62692]: DEBUG nova.virt.hardware [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1063.629492] env[62692]: DEBUG nova.virt.hardware [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1063.629680] env[62692]: DEBUG nova.virt.hardware [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1063.629834] env[62692]: DEBUG nova.virt.hardware [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1063.629989] env[62692]: DEBUG nova.virt.hardware [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1063.630250] env[62692]: DEBUG nova.virt.hardware [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1063.630434] env[62692]: DEBUG nova.virt.hardware [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1063.630612] env[62692]: DEBUG nova.virt.hardware [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1063.630779] env[62692]: DEBUG nova.virt.hardware [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1063.630955] env[62692]: DEBUG nova.virt.hardware [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1063.631812] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6187c957-70b7-4197-9091-2689a537be44 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.640718] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-110a2f0f-4555-43d9-82a7-969b97d9a46c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.655151] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Instance VIF info [] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1063.660748] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Creating folder: Project (d11788c7c2fd48e1a0552fb57bdab468). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1063.661073] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4b46f2d0-ce07-4f3c-8986-2aaecd97bbf4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.672227] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Created folder: Project (d11788c7c2fd48e1a0552fb57bdab468) in parent group-v248868. [ 1063.672441] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Creating folder: Instances. Parent ref: group-v249033. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1063.672723] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ae48b356-99b7-487f-b60f-7eb009cffaef {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.681716] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Created folder: Instances in parent group-v249033. [ 1063.681963] env[62692]: DEBUG oslo.service.loopingcall [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1063.682205] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1063.682440] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-50b4bda2-17ce-40c5-a694-c61414b31ee5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.702032] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1063.702032] env[62692]: value = "task-1141719" [ 1063.702032] env[62692]: _type = "Task" [ 1063.702032] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.707606] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141719, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.801858] env[62692]: DEBUG nova.compute.manager [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1063.855100] env[62692]: DEBUG oslo_concurrency.lockutils [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.278s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.855311] env[62692]: INFO nova.compute.manager [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Migrating [ 1063.861851] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.258s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.862098] env[62692]: DEBUG nova.objects.instance [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lazy-loading 'resources' on Instance uuid bb547773-d176-4c8e-a0fa-a374d5050b1b {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1063.877485] env[62692]: DEBUG oslo_concurrency.lockutils [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Releasing lock "refresh_cache-90932f05-fad3-4c6d-87ae-ab059351b0be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.877715] env[62692]: DEBUG nova.compute.manager [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Instance network_info: |[{"id": "21e76034-6c0f-4938-afda-0d005cf45759", "address": "fa:16:3e:2d:c0:4a", "network": {"id": "77695f27-3027-4a92-8b9b-0a90cdc53f62", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-337681783-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697b4281872f4547ad3d051e644b95ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21e76034-6c", "ovs_interfaceid": "21e76034-6c0f-4938-afda-0d005cf45759", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1063.878616] env[62692]: DEBUG oslo_concurrency.lockutils [req-bcd9e6b8-4521-4074-82a1-3b7227e9a902 req-1a93ca3d-3d7d-4daf-a857-26c304da383b service nova] Releasing lock "refresh_cache-1b753882-9118-479c-aa3e-cbb8bdf2e086" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.879193] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2d:c0:4a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8e7f6f41-f4eb-4832-a390-730fca1cf717', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '21e76034-6c0f-4938-afda-0d005cf45759', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1063.887140] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Creating folder: Project (697b4281872f4547ad3d051e644b95ce). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1063.888420] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a69849c1-cfa5-4714-bfb0-9c1c83794f9a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.901857] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Created folder: Project (697b4281872f4547ad3d051e644b95ce) in parent group-v248868. [ 1063.902088] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Creating folder: Instances. Parent ref: group-v249036. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1063.902340] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-36545204-ec1f-4325-978e-ef1b878d2f1e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.913170] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Created folder: Instances in parent group-v249036. [ 1063.913429] env[62692]: DEBUG oslo.service.loopingcall [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1063.913620] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1063.914161] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2c4cce60-fb67-444a-9bd2-cdfc5bb57a6d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.931509] env[62692]: DEBUG nova.compute.manager [req-1a85e151-6c27-41a0-8e7c-eaf76d546e8c req-d0c0c8ab-110b-4558-a9b6-890ff0cb08a8 service nova] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Received event network-vif-plugged-21e76034-6c0f-4938-afda-0d005cf45759 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1063.932261] env[62692]: DEBUG oslo_concurrency.lockutils [req-1a85e151-6c27-41a0-8e7c-eaf76d546e8c req-d0c0c8ab-110b-4558-a9b6-890ff0cb08a8 service nova] Acquiring lock "90932f05-fad3-4c6d-87ae-ab059351b0be-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.932261] env[62692]: DEBUG oslo_concurrency.lockutils [req-1a85e151-6c27-41a0-8e7c-eaf76d546e8c req-d0c0c8ab-110b-4558-a9b6-890ff0cb08a8 service nova] Lock "90932f05-fad3-4c6d-87ae-ab059351b0be-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.932261] env[62692]: DEBUG oslo_concurrency.lockutils [req-1a85e151-6c27-41a0-8e7c-eaf76d546e8c req-d0c0c8ab-110b-4558-a9b6-890ff0cb08a8 service nova] Lock "90932f05-fad3-4c6d-87ae-ab059351b0be-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.932261] env[62692]: DEBUG nova.compute.manager [req-1a85e151-6c27-41a0-8e7c-eaf76d546e8c req-d0c0c8ab-110b-4558-a9b6-890ff0cb08a8 service nova] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] No waiting events found dispatching network-vif-plugged-21e76034-6c0f-4938-afda-0d005cf45759 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1063.932440] env[62692]: WARNING nova.compute.manager [req-1a85e151-6c27-41a0-8e7c-eaf76d546e8c req-d0c0c8ab-110b-4558-a9b6-890ff0cb08a8 service nova] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Received unexpected event network-vif-plugged-21e76034-6c0f-4938-afda-0d005cf45759 for instance with vm_state building and task_state spawning. [ 1063.932648] env[62692]: DEBUG nova.compute.manager [req-1a85e151-6c27-41a0-8e7c-eaf76d546e8c req-d0c0c8ab-110b-4558-a9b6-890ff0cb08a8 service nova] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Received event network-changed-21e76034-6c0f-4938-afda-0d005cf45759 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1063.932748] env[62692]: DEBUG nova.compute.manager [req-1a85e151-6c27-41a0-8e7c-eaf76d546e8c req-d0c0c8ab-110b-4558-a9b6-890ff0cb08a8 service nova] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Refreshing instance network info cache due to event network-changed-21e76034-6c0f-4938-afda-0d005cf45759. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1063.932936] env[62692]: DEBUG oslo_concurrency.lockutils [req-1a85e151-6c27-41a0-8e7c-eaf76d546e8c req-d0c0c8ab-110b-4558-a9b6-890ff0cb08a8 service nova] Acquiring lock "refresh_cache-90932f05-fad3-4c6d-87ae-ab059351b0be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.933086] env[62692]: DEBUG oslo_concurrency.lockutils [req-1a85e151-6c27-41a0-8e7c-eaf76d546e8c req-d0c0c8ab-110b-4558-a9b6-890ff0cb08a8 service nova] Acquired lock "refresh_cache-90932f05-fad3-4c6d-87ae-ab059351b0be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.933251] env[62692]: DEBUG nova.network.neutron [req-1a85e151-6c27-41a0-8e7c-eaf76d546e8c req-d0c0c8ab-110b-4558-a9b6-890ff0cb08a8 service nova] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Refreshing network info cache for port 21e76034-6c0f-4938-afda-0d005cf45759 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1063.939848] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1063.939848] env[62692]: value = "task-1141722" [ 1063.939848] env[62692]: _type = "Task" [ 1063.939848] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.949175] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141722, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.044199] env[62692]: DEBUG oslo_vmware.api [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141716, 'name': ReconfigVM_Task, 'duration_secs': 0.56255} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.044504] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Reconfigured VM instance instance-0000005d to attach disk [datastore2] ca56c2ec-95fc-44fc-856f-e395bda1703c/ca56c2ec-95fc-44fc-856f-e395bda1703c.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1064.045142] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d44f1576-6408-4d5c-b78b-19ae7b4b9ed0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.052164] env[62692]: DEBUG oslo_vmware.api [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1064.052164] env[62692]: value = "task-1141723" [ 1064.052164] env[62692]: _type = "Task" [ 1064.052164] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.061877] env[62692]: DEBUG oslo_vmware.api [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141723, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.210279] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141719, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.323429] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.375094] env[62692]: DEBUG oslo_concurrency.lockutils [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "refresh_cache-9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.375315] env[62692]: DEBUG oslo_concurrency.lockutils [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired lock "refresh_cache-9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.375505] env[62692]: DEBUG nova.network.neutron [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1064.453671] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141722, 'name': CreateVM_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.563477] env[62692]: DEBUG oslo_vmware.api [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141723, 'name': Rename_Task, 'duration_secs': 0.152135} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.563765] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1064.564026] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-74dbd6ec-8a52-4532-8671-f1247df48c4d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.566720] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-350f4b4a-61b1-45f9-9346-f4c29b66a0fa {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.573833] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48388bc8-96bf-4d73-8ad5-df30dea46cfd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.577934] env[62692]: DEBUG oslo_vmware.api [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1064.577934] env[62692]: value = "task-1141724" [ 1064.577934] env[62692]: _type = "Task" [ 1064.577934] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.610472] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b10f76a7-7ce6-41dd-8d86-c6df676eef56 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.617368] env[62692]: DEBUG oslo_vmware.api [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141724, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.622570] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64c46739-5858-4609-94d1-33e5765d6ddd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.636031] env[62692]: DEBUG nova.compute.provider_tree [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1064.689181] env[62692]: DEBUG nova.network.neutron [req-1a85e151-6c27-41a0-8e7c-eaf76d546e8c req-d0c0c8ab-110b-4558-a9b6-890ff0cb08a8 service nova] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Updated VIF entry in instance network info cache for port 21e76034-6c0f-4938-afda-0d005cf45759. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1064.689608] env[62692]: DEBUG nova.network.neutron [req-1a85e151-6c27-41a0-8e7c-eaf76d546e8c req-d0c0c8ab-110b-4558-a9b6-890ff0cb08a8 service nova] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Updating instance_info_cache with network_info: [{"id": "21e76034-6c0f-4938-afda-0d005cf45759", "address": "fa:16:3e:2d:c0:4a", "network": {"id": "77695f27-3027-4a92-8b9b-0a90cdc53f62", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-337681783-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697b4281872f4547ad3d051e644b95ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21e76034-6c", "ovs_interfaceid": "21e76034-6c0f-4938-afda-0d005cf45759", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1064.712543] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141719, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.953607] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141722, 'name': CreateVM_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.091260] env[62692]: DEBUG oslo_vmware.api [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141724, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.140432] env[62692]: DEBUG nova.scheduler.client.report [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1065.186821] env[62692]: DEBUG nova.network.neutron [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Updating instance_info_cache with network_info: [{"id": "d5ff0422-e9b3-455c-bd28-422096794288", "address": "fa:16:3e:46:38:ff", "network": {"id": "f6352999-f943-431d-b54a-c9dc1d810a85", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-121037111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9cb8ae08824245a7911de7455e7fcb68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c842425c-544e-4ce2-9657-512723bd318e", "external-id": "nsx-vlan-transportzone-80", "segmentation_id": 80, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5ff0422-e9", "ovs_interfaceid": "d5ff0422-e9b3-455c-bd28-422096794288", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.192539] env[62692]: DEBUG oslo_concurrency.lockutils [req-1a85e151-6c27-41a0-8e7c-eaf76d546e8c req-d0c0c8ab-110b-4558-a9b6-890ff0cb08a8 service nova] Releasing lock "refresh_cache-90932f05-fad3-4c6d-87ae-ab059351b0be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.215952] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141719, 'name': CreateVM_Task, 'duration_secs': 1.341559} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.216580] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1065.217076] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.217252] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.217588] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1065.218272] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cbd0d1fe-f4eb-4d31-a042-58992cc1d695 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.226360] env[62692]: DEBUG oslo_vmware.api [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Waiting for the task: (returnval){ [ 1065.226360] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52e08696-fa8c-d2cc-851a-ba22fa0b3cc7" [ 1065.226360] env[62692]: _type = "Task" [ 1065.226360] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.236393] env[62692]: DEBUG oslo_vmware.api [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52e08696-fa8c-d2cc-851a-ba22fa0b3cc7, 'name': SearchDatastore_Task, 'duration_secs': 0.009191} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.236722] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.237019] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1065.237303] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.237463] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.237645] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1065.237934] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-22638786-6042-42b8-8d13-0847d4d60f52 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.247241] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1065.247451] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1065.248211] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57aa094e-c35f-4af4-af29-89ecaa4f3bf9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.254971] env[62692]: DEBUG oslo_vmware.api [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Waiting for the task: (returnval){ [ 1065.254971] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5298ed40-84f6-ca9d-79b6-75d525d37fcd" [ 1065.254971] env[62692]: _type = "Task" [ 1065.254971] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.265664] env[62692]: DEBUG oslo_vmware.api [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5298ed40-84f6-ca9d-79b6-75d525d37fcd, 'name': SearchDatastore_Task} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.265919] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-58545688-4152-49ef-81a5-05a759c650f8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.273284] env[62692]: DEBUG oslo_vmware.api [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Waiting for the task: (returnval){ [ 1065.273284] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]520c683d-4321-1075-a28c-d30728d4b310" [ 1065.273284] env[62692]: _type = "Task" [ 1065.273284] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.281076] env[62692]: DEBUG oslo_vmware.api [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]520c683d-4321-1075-a28c-d30728d4b310, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.453914] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141722, 'name': CreateVM_Task, 'duration_secs': 1.280872} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.454227] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1065.454957] env[62692]: DEBUG oslo_concurrency.lockutils [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.454957] env[62692]: DEBUG oslo_concurrency.lockutils [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.455495] env[62692]: DEBUG oslo_concurrency.lockutils [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1065.455565] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ea154cb-e85d-4491-b3fc-507f0df6e53a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.461520] env[62692]: DEBUG oslo_vmware.api [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1065.461520] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52b280ad-0bd5-e1c4-8c84-00b86d30f5c4" [ 1065.461520] env[62692]: _type = "Task" [ 1065.461520] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.470923] env[62692]: DEBUG oslo_vmware.api [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52b280ad-0bd5-e1c4-8c84-00b86d30f5c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.590719] env[62692]: DEBUG oslo_vmware.api [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141724, 'name': PowerOnVM_Task, 'duration_secs': 0.581532} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.591058] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1065.591288] env[62692]: INFO nova.compute.manager [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Took 8.61 seconds to spawn the instance on the hypervisor. [ 1065.591482] env[62692]: DEBUG nova.compute.manager [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1065.592556] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d20b1360-6645-4307-be55-93314589866d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.646829] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.784s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.648647] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.173s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.650404] env[62692]: INFO nova.compute.claims [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1065.670995] env[62692]: INFO nova.scheduler.client.report [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Deleted allocations for instance bb547773-d176-4c8e-a0fa-a374d5050b1b [ 1065.691558] env[62692]: DEBUG oslo_concurrency.lockutils [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Releasing lock "refresh_cache-9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.784827] env[62692]: DEBUG oslo_vmware.api [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]520c683d-4321-1075-a28c-d30728d4b310, 'name': SearchDatastore_Task, 'duration_secs': 0.009025} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.785278] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.786032] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] e086d79d-29e6-41bf-a139-0b680ca0021d/e086d79d-29e6-41bf-a139-0b680ca0021d.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1065.786147] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4cddf3f3-f685-465a-82c5-a8866df035e9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.797026] env[62692]: DEBUG oslo_vmware.api [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Waiting for the task: (returnval){ [ 1065.797026] env[62692]: value = "task-1141725" [ 1065.797026] env[62692]: _type = "Task" [ 1065.797026] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.808037] env[62692]: DEBUG oslo_vmware.api [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141725, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.973893] env[62692]: DEBUG oslo_vmware.api [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52b280ad-0bd5-e1c4-8c84-00b86d30f5c4, 'name': SearchDatastore_Task, 'duration_secs': 0.010668} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.974329] env[62692]: DEBUG oslo_concurrency.lockutils [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.974546] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1065.974828] env[62692]: DEBUG oslo_concurrency.lockutils [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.975015] env[62692]: DEBUG oslo_concurrency.lockutils [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.975275] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1065.975575] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1a8f5156-ee80-4ed1-b48b-044e477289d7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.986093] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1065.986387] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1065.987289] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a34670e-aa32-4a56-8f2f-20f40c61c062 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.994638] env[62692]: DEBUG oslo_vmware.api [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1065.994638] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5208d558-c4c7-414b-42de-fc31cdb82674" [ 1065.994638] env[62692]: _type = "Task" [ 1065.994638] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.007832] env[62692]: DEBUG oslo_vmware.api [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5208d558-c4c7-414b-42de-fc31cdb82674, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.114438] env[62692]: INFO nova.compute.manager [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Took 26.79 seconds to build instance. [ 1066.179438] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6739f11d-d005-408c-a032-c91e69bbcc36 tempest-ServerRescueNegativeTestJSON-1995122140 tempest-ServerRescueNegativeTestJSON-1995122140-project-member] Lock "bb547773-d176-4c8e-a0fa-a374d5050b1b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.699s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.312459] env[62692]: DEBUG oslo_vmware.api [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141725, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.507969] env[62692]: DEBUG oslo_vmware.api [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5208d558-c4c7-414b-42de-fc31cdb82674, 'name': SearchDatastore_Task, 'duration_secs': 0.076205} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.509116] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15395a5e-a5d4-4b8f-a8af-95e4058f97c2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.516030] env[62692]: DEBUG oslo_vmware.api [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1066.516030] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52cb73cb-dc76-f085-e3a4-89ea8132046b" [ 1066.516030] env[62692]: _type = "Task" [ 1066.516030] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.525489] env[62692]: DEBUG oslo_vmware.api [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52cb73cb-dc76-f085-e3a4-89ea8132046b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.617415] env[62692]: DEBUG oslo_concurrency.lockutils [None req-89db3d82-855c-4aa3-bfc9-19853735520b tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "ca56c2ec-95fc-44fc-856f-e395bda1703c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.307s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.813772] env[62692]: DEBUG oslo_vmware.api [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141725, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.61562} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.813772] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] e086d79d-29e6-41bf-a139-0b680ca0021d/e086d79d-29e6-41bf-a139-0b680ca0021d.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1066.813772] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1066.813772] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a66d1ba1-74e3-4f48-a5e0-2cc1bbea29be {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.825502] env[62692]: DEBUG oslo_vmware.api [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Waiting for the task: (returnval){ [ 1066.825502] env[62692]: value = "task-1141726" [ 1066.825502] env[62692]: _type = "Task" [ 1066.825502] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.835173] env[62692]: DEBUG oslo_vmware.api [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141726, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.844717] env[62692]: DEBUG oslo_concurrency.lockutils [None req-da1a7a7c-deb9-4f3f-9bd1-0bb2cffd2437 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "ca56c2ec-95fc-44fc-856f-e395bda1703c" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.845010] env[62692]: DEBUG oslo_concurrency.lockutils [None req-da1a7a7c-deb9-4f3f-9bd1-0bb2cffd2437 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "ca56c2ec-95fc-44fc-856f-e395bda1703c" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.845210] env[62692]: DEBUG nova.compute.manager [None req-da1a7a7c-deb9-4f3f-9bd1-0bb2cffd2437 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1066.846249] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26a7035a-0b69-4274-b12d-7db84d97c6ea {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.858356] env[62692]: DEBUG nova.compute.manager [None req-da1a7a7c-deb9-4f3f-9bd1-0bb2cffd2437 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62692) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1066.858991] env[62692]: DEBUG nova.objects.instance [None req-da1a7a7c-deb9-4f3f-9bd1-0bb2cffd2437 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lazy-loading 'flavor' on Instance uuid ca56c2ec-95fc-44fc-856f-e395bda1703c {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1066.931517] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5aba418-f65a-4433-81bc-a5862c0780a9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.943053] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92992e60-d104-434e-9277-769f2a1c7eb0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.980398] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54f3d036-77c3-495c-bfe0-cba0e7b7575b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.989425] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0469afb-8557-4a6e-9a66-529fe497faf9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.005129] env[62692]: DEBUG nova.compute.provider_tree [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1067.030168] env[62692]: DEBUG oslo_vmware.api [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52cb73cb-dc76-f085-e3a4-89ea8132046b, 'name': SearchDatastore_Task, 'duration_secs': 0.095691} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.030397] env[62692]: DEBUG oslo_concurrency.lockutils [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.030596] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 90932f05-fad3-4c6d-87ae-ab059351b0be/90932f05-fad3-4c6d-87ae-ab059351b0be.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1067.030873] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5b469b93-4d89-4ab4-8e85-531885864be4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.039411] env[62692]: DEBUG oslo_vmware.api [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1067.039411] env[62692]: value = "task-1141727" [ 1067.039411] env[62692]: _type = "Task" [ 1067.039411] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.049251] env[62692]: DEBUG oslo_vmware.api [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141727, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.211756] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5de98e3-6428-4273-b98f-32e6e4a867e6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.233191] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Updating instance '9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da' progress to 0 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1067.341820] env[62692]: DEBUG oslo_vmware.api [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141726, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.264743} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.342834] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1067.343954] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4930b776-d45e-4cad-bf02-df2bebdee689 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.369830] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] e086d79d-29e6-41bf-a139-0b680ca0021d/e086d79d-29e6-41bf-a139-0b680ca0021d.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1067.372238] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4a322d70-0ef4-4b5e-9918-6bbbaa2ed50b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.387804] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-da1a7a7c-deb9-4f3f-9bd1-0bb2cffd2437 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1067.388192] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8fc29579-741c-418f-95bd-60db2828eff6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.396926] env[62692]: DEBUG oslo_vmware.api [None req-da1a7a7c-deb9-4f3f-9bd1-0bb2cffd2437 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1067.396926] env[62692]: value = "task-1141728" [ 1067.396926] env[62692]: _type = "Task" [ 1067.396926] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.398925] env[62692]: DEBUG oslo_vmware.api [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Waiting for the task: (returnval){ [ 1067.398925] env[62692]: value = "task-1141729" [ 1067.398925] env[62692]: _type = "Task" [ 1067.398925] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.414111] env[62692]: DEBUG oslo_vmware.api [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141729, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.417878] env[62692]: DEBUG oslo_vmware.api [None req-da1a7a7c-deb9-4f3f-9bd1-0bb2cffd2437 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141728, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.509704] env[62692]: DEBUG nova.scheduler.client.report [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1067.552832] env[62692]: DEBUG oslo_vmware.api [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141727, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.739226] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1067.739559] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f180adc9-789e-494d-bfff-9dba041f34ac {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.748033] env[62692]: DEBUG oslo_vmware.api [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1067.748033] env[62692]: value = "task-1141730" [ 1067.748033] env[62692]: _type = "Task" [ 1067.748033] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.765981] env[62692]: DEBUG oslo_vmware.api [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141730, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.914261] env[62692]: DEBUG oslo_vmware.api [None req-da1a7a7c-deb9-4f3f-9bd1-0bb2cffd2437 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141728, 'name': PowerOffVM_Task, 'duration_secs': 0.287081} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.921404] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-da1a7a7c-deb9-4f3f-9bd1-0bb2cffd2437 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1067.921404] env[62692]: DEBUG nova.compute.manager [None req-da1a7a7c-deb9-4f3f-9bd1-0bb2cffd2437 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1067.921404] env[62692]: DEBUG oslo_vmware.api [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141729, 'name': ReconfigVM_Task, 'duration_secs': 0.452057} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.921404] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c87338f-4f5e-4824-9690-b778bede1c50 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.922611] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Reconfigured VM instance instance-0000005f to attach disk [datastore2] e086d79d-29e6-41bf-a139-0b680ca0021d/e086d79d-29e6-41bf-a139-0b680ca0021d.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1067.923177] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b343321e-197a-4e55-a032-89dcefc88602 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.936137] env[62692]: DEBUG oslo_vmware.api [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Waiting for the task: (returnval){ [ 1067.936137] env[62692]: value = "task-1141731" [ 1067.936137] env[62692]: _type = "Task" [ 1067.936137] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.946303] env[62692]: DEBUG oslo_vmware.api [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141731, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.017655] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.369s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.018221] env[62692]: DEBUG nova.compute.manager [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1068.022164] env[62692]: DEBUG oslo_concurrency.lockutils [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.791s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.022864] env[62692]: DEBUG nova.objects.instance [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Lazy-loading 'resources' on Instance uuid af260f9a-4e9b-4af1-90da-c619c4755eca {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1068.052749] env[62692]: DEBUG oslo_vmware.api [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141727, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.533896} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.053028] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 90932f05-fad3-4c6d-87ae-ab059351b0be/90932f05-fad3-4c6d-87ae-ab059351b0be.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1068.053246] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1068.053567] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bb1b29af-4e06-41f6-80f0-7f022ba475b8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.062787] env[62692]: DEBUG oslo_vmware.api [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1068.062787] env[62692]: value = "task-1141732" [ 1068.062787] env[62692]: _type = "Task" [ 1068.062787] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.076323] env[62692]: DEBUG oslo_vmware.api [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141732, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.258699] env[62692]: DEBUG oslo_vmware.api [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141730, 'name': PowerOffVM_Task, 'duration_secs': 0.203407} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.258699] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1068.258831] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Updating instance '9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da' progress to 17 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1068.444307] env[62692]: DEBUG oslo_concurrency.lockutils [None req-da1a7a7c-deb9-4f3f-9bd1-0bb2cffd2437 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "ca56c2ec-95fc-44fc-856f-e395bda1703c" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.598s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.450589] env[62692]: DEBUG oslo_vmware.api [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141731, 'name': Rename_Task, 'duration_secs': 0.155654} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.451288] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1068.451288] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6c669bb0-3a2f-409a-abbc-c9fd18fe5a52 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.459748] env[62692]: DEBUG oslo_vmware.api [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Waiting for the task: (returnval){ [ 1068.459748] env[62692]: value = "task-1141733" [ 1068.459748] env[62692]: _type = "Task" [ 1068.459748] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.473584] env[62692]: DEBUG oslo_vmware.api [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141733, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.527036] env[62692]: DEBUG nova.compute.utils [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1068.528870] env[62692]: DEBUG nova.compute.manager [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1068.528870] env[62692]: DEBUG nova.network.neutron [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1068.581208] env[62692]: DEBUG oslo_vmware.api [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141732, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081836} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.582249] env[62692]: DEBUG nova.policy [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b0c1314a9aed42e5b54a33de96f3db1f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '71b00c566cea4b21811ea1e1ef5ba0f3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1068.584027] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1068.585499] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f39a5a3-cbc7-45ad-8fa7-41b0dbfd05b1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.620137] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 90932f05-fad3-4c6d-87ae-ab059351b0be/90932f05-fad3-4c6d-87ae-ab059351b0be.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1068.620137] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-48e937f7-fb76-4011-af05-60f90fa3620a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.645129] env[62692]: DEBUG oslo_vmware.api [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1068.645129] env[62692]: value = "task-1141734" [ 1068.645129] env[62692]: _type = "Task" [ 1068.645129] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.655228] env[62692]: DEBUG oslo_vmware.api [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141734, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.766581] env[62692]: DEBUG nova.virt.hardware [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1068.766908] env[62692]: DEBUG nova.virt.hardware [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1068.767249] env[62692]: DEBUG nova.virt.hardware [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1068.767359] env[62692]: DEBUG nova.virt.hardware [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1068.767430] env[62692]: DEBUG nova.virt.hardware [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1068.768028] env[62692]: DEBUG nova.virt.hardware [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1068.768028] env[62692]: DEBUG nova.virt.hardware [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1068.768028] env[62692]: DEBUG nova.virt.hardware [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1068.770667] env[62692]: DEBUG nova.virt.hardware [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1068.770667] env[62692]: DEBUG nova.virt.hardware [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1068.770667] env[62692]: DEBUG nova.virt.hardware [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1068.777087] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9aa73ec9-d327-4f9e-afab-2bcd88ee6baa {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.807543] env[62692]: DEBUG oslo_vmware.api [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1068.807543] env[62692]: value = "task-1141735" [ 1068.807543] env[62692]: _type = "Task" [ 1068.807543] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.818218] env[62692]: DEBUG oslo_vmware.api [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141735, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.878906] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7697aca-422a-4d76-8b1c-deef8f5ff9b3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.887937] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53ce404e-ccb4-4c0b-98ed-fe758f112afb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.929405] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d76d75f4-83d1-4498-9df8-5137a0c2ba9f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.939449] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94bd1ab1-e2be-43a7-bccd-844fae619e08 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.957217] env[62692]: DEBUG nova.compute.provider_tree [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1068.970531] env[62692]: DEBUG oslo_vmware.api [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141733, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.013744] env[62692]: DEBUG nova.network.neutron [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Successfully created port: c3aed689-318a-45c3-a236-6ca6b1bace9b {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1069.035798] env[62692]: DEBUG nova.compute.manager [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1069.158143] env[62692]: DEBUG oslo_vmware.api [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141734, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.322209] env[62692]: DEBUG oslo_vmware.api [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141735, 'name': ReconfigVM_Task, 'duration_secs': 0.301505} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.322544] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Updating instance '9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da' progress to 33 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1069.460817] env[62692]: DEBUG nova.scheduler.client.report [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1069.479988] env[62692]: DEBUG oslo_vmware.api [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141733, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.628820] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "ca56c2ec-95fc-44fc-856f-e395bda1703c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.629242] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "ca56c2ec-95fc-44fc-856f-e395bda1703c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.629353] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "ca56c2ec-95fc-44fc-856f-e395bda1703c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.629570] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "ca56c2ec-95fc-44fc-856f-e395bda1703c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.629713] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "ca56c2ec-95fc-44fc-856f-e395bda1703c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.632365] env[62692]: INFO nova.compute.manager [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Terminating instance [ 1069.636179] env[62692]: DEBUG nova.compute.manager [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1069.636179] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1069.636179] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64b81192-9304-4182-b1e6-60e8b337ca22 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.644641] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1069.645000] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f6cc5dec-abaf-4fed-9ea8-653325c05076 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.656220] env[62692]: DEBUG oslo_vmware.api [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141734, 'name': ReconfigVM_Task, 'duration_secs': 0.810079} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.656501] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 90932f05-fad3-4c6d-87ae-ab059351b0be/90932f05-fad3-4c6d-87ae-ab059351b0be.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1069.657250] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cfa43836-86fa-456f-bc5f-c0a98e86687c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.664149] env[62692]: DEBUG oslo_vmware.api [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1069.664149] env[62692]: value = "task-1141737" [ 1069.664149] env[62692]: _type = "Task" [ 1069.664149] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.727882] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1069.728251] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1069.728476] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Deleting the datastore file [datastore2] ca56c2ec-95fc-44fc-856f-e395bda1703c {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1069.728766] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4a2084b1-a32e-49f4-9708-401335557aed {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.737761] env[62692]: DEBUG oslo_vmware.api [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1069.737761] env[62692]: value = "task-1141738" [ 1069.737761] env[62692]: _type = "Task" [ 1069.737761] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.746405] env[62692]: DEBUG oslo_vmware.api [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141738, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.829997] env[62692]: DEBUG nova.virt.hardware [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1069.830086] env[62692]: DEBUG nova.virt.hardware [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1069.830342] env[62692]: DEBUG nova.virt.hardware [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1069.830592] env[62692]: DEBUG nova.virt.hardware [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1069.830774] env[62692]: DEBUG nova.virt.hardware [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1069.830967] env[62692]: DEBUG nova.virt.hardware [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1069.831249] env[62692]: DEBUG nova.virt.hardware [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1069.831467] env[62692]: DEBUG nova.virt.hardware [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1069.831678] env[62692]: DEBUG nova.virt.hardware [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1069.831855] env[62692]: DEBUG nova.virt.hardware [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1069.832094] env[62692]: DEBUG nova.virt.hardware [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1069.837269] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Reconfiguring VM instance instance-0000005a to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1069.837575] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e4ad3163-2142-43a7-843e-0dabe70e3991 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.857275] env[62692]: DEBUG oslo_vmware.api [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1069.857275] env[62692]: value = "task-1141739" [ 1069.857275] env[62692]: _type = "Task" [ 1069.857275] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.866416] env[62692]: DEBUG oslo_vmware.api [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141739, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.971048] env[62692]: DEBUG oslo_concurrency.lockutils [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.948s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.973361] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.532s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.973628] env[62692]: DEBUG nova.objects.instance [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lazy-loading 'resources' on Instance uuid a1951dba-ec5a-4849-9e22-37fba08ec60c {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1069.984899] env[62692]: DEBUG oslo_vmware.api [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141733, 'name': PowerOnVM_Task, 'duration_secs': 1.303511} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.985077] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1069.985308] env[62692]: INFO nova.compute.manager [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Took 6.38 seconds to spawn the instance on the hypervisor. [ 1069.985757] env[62692]: DEBUG nova.compute.manager [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1069.987069] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-018033c5-f064-40fb-b00c-a13520f494f1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.999500] env[62692]: INFO nova.scheduler.client.report [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Deleted allocations for instance af260f9a-4e9b-4af1-90da-c619c4755eca [ 1070.055296] env[62692]: DEBUG nova.compute.manager [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1070.075876] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Acquiring lock "f27e3a4f-d109-472b-a348-5099bdc0e7c8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.076123] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Lock "f27e3a4f-d109-472b-a348-5099bdc0e7c8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.087710] env[62692]: DEBUG nova.virt.hardware [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1070.087942] env[62692]: DEBUG nova.virt.hardware [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1070.088121] env[62692]: DEBUG nova.virt.hardware [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1070.088321] env[62692]: DEBUG nova.virt.hardware [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1070.088497] env[62692]: DEBUG nova.virt.hardware [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1070.088665] env[62692]: DEBUG nova.virt.hardware [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1070.088855] env[62692]: DEBUG nova.virt.hardware [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1070.089021] env[62692]: DEBUG nova.virt.hardware [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1070.089195] env[62692]: DEBUG nova.virt.hardware [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1070.089358] env[62692]: DEBUG nova.virt.hardware [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1070.089531] env[62692]: DEBUG nova.virt.hardware [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1070.090438] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d04e039c-c78d-489c-b09e-c915248e5483 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.100019] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15832539-c307-42ef-96f1-9dbe7b73a72f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.174133] env[62692]: DEBUG oslo_vmware.api [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141737, 'name': Rename_Task, 'duration_secs': 0.175566} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.174423] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1070.174666] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8afafd56-0414-48d7-a71e-924f288e3465 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.183181] env[62692]: DEBUG oslo_vmware.api [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1070.183181] env[62692]: value = "task-1141740" [ 1070.183181] env[62692]: _type = "Task" [ 1070.183181] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.190043] env[62692]: DEBUG oslo_vmware.api [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141740, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.250769] env[62692]: DEBUG oslo_vmware.api [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141738, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.203617} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.250769] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1070.250769] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1070.250769] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1070.250769] env[62692]: INFO nova.compute.manager [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Took 0.61 seconds to destroy the instance on the hypervisor. [ 1070.250769] env[62692]: DEBUG oslo.service.loopingcall [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1070.250769] env[62692]: DEBUG nova.compute.manager [-] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1070.250769] env[62692]: DEBUG nova.network.neutron [-] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1070.367917] env[62692]: DEBUG oslo_vmware.api [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141739, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.512758] env[62692]: INFO nova.compute.manager [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Took 30.29 seconds to build instance. [ 1070.515323] env[62692]: DEBUG oslo_concurrency.lockutils [None req-089ef37d-6d9c-4a7b-8a59-5f6a32447955 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Lock "af260f9a-4e9b-4af1-90da-c619c4755eca" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.286s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.520285] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2650f739-3dd3-49fb-ae92-85115406f198 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Lock "af260f9a-4e9b-4af1-90da-c619c4755eca" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 8.675s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.520285] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2650f739-3dd3-49fb-ae92-85115406f198 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquiring lock "af260f9a-4e9b-4af1-90da-c619c4755eca-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.520285] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2650f739-3dd3-49fb-ae92-85115406f198 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Lock "af260f9a-4e9b-4af1-90da-c619c4755eca-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.520285] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2650f739-3dd3-49fb-ae92-85115406f198 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Lock "af260f9a-4e9b-4af1-90da-c619c4755eca-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.525781] env[62692]: INFO nova.compute.manager [None req-2650f739-3dd3-49fb-ae92-85115406f198 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Terminating instance [ 1070.528187] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2650f739-3dd3-49fb-ae92-85115406f198 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquiring lock "refresh_cache-af260f9a-4e9b-4af1-90da-c619c4755eca" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.528362] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2650f739-3dd3-49fb-ae92-85115406f198 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquired lock "refresh_cache-af260f9a-4e9b-4af1-90da-c619c4755eca" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.528800] env[62692]: DEBUG nova.network.neutron [None req-2650f739-3dd3-49fb-ae92-85115406f198 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1070.551895] env[62692]: DEBUG nova.compute.manager [req-f5a02caa-43a1-4253-a795-f6399a5f8cdc req-fff02e25-7061-42b9-a982-480a6cec594b service nova] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Received event network-vif-deleted-898dbc93-ff46-4a9e-9214-060a76c66507 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1070.552130] env[62692]: INFO nova.compute.manager [req-f5a02caa-43a1-4253-a795-f6399a5f8cdc req-fff02e25-7061-42b9-a982-480a6cec594b service nova] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Neutron deleted interface 898dbc93-ff46-4a9e-9214-060a76c66507; detaching it from the instance and deleting it from the info cache [ 1070.552334] env[62692]: DEBUG nova.network.neutron [req-f5a02caa-43a1-4253-a795-f6399a5f8cdc req-fff02e25-7061-42b9-a982-480a6cec594b service nova] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.580473] env[62692]: DEBUG nova.compute.manager [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1070.695449] env[62692]: DEBUG oslo_vmware.api [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141740, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.722094] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a21474c-434a-46c5-8c42-7812200ec32c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.732212] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69cd7648-7778-4a8f-b067-3cafca68dc36 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.768562] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fb7380f-f7fb-4c7d-83c9-f6fed4b82af7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.772279] env[62692]: DEBUG nova.compute.manager [req-c33f5a1d-0c75-45ca-9ffd-6851189e3b47 req-9aab398a-ca25-41b3-be04-a9a0f0a312ac service nova] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Received event network-vif-plugged-c3aed689-318a-45c3-a236-6ca6b1bace9b {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1070.772785] env[62692]: DEBUG oslo_concurrency.lockutils [req-c33f5a1d-0c75-45ca-9ffd-6851189e3b47 req-9aab398a-ca25-41b3-be04-a9a0f0a312ac service nova] Acquiring lock "a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.772937] env[62692]: DEBUG oslo_concurrency.lockutils [req-c33f5a1d-0c75-45ca-9ffd-6851189e3b47 req-9aab398a-ca25-41b3-be04-a9a0f0a312ac service nova] Lock "a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.773180] env[62692]: DEBUG oslo_concurrency.lockutils [req-c33f5a1d-0c75-45ca-9ffd-6851189e3b47 req-9aab398a-ca25-41b3-be04-a9a0f0a312ac service nova] Lock "a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.773356] env[62692]: DEBUG nova.compute.manager [req-c33f5a1d-0c75-45ca-9ffd-6851189e3b47 req-9aab398a-ca25-41b3-be04-a9a0f0a312ac service nova] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] No waiting events found dispatching network-vif-plugged-c3aed689-318a-45c3-a236-6ca6b1bace9b {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1070.773865] env[62692]: WARNING nova.compute.manager [req-c33f5a1d-0c75-45ca-9ffd-6851189e3b47 req-9aab398a-ca25-41b3-be04-a9a0f0a312ac service nova] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Received unexpected event network-vif-plugged-c3aed689-318a-45c3-a236-6ca6b1bace9b for instance with vm_state building and task_state spawning. [ 1070.784931] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e90a7a2-d8cd-4215-a3b9-efa912597cbc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.801461] env[62692]: DEBUG nova.compute.provider_tree [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1070.822251] env[62692]: DEBUG nova.network.neutron [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Successfully updated port: c3aed689-318a-45c3-a236-6ca6b1bace9b {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1070.870734] env[62692]: DEBUG oslo_vmware.api [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141739, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.935117] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1070.935117] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1070.962721] env[62692]: DEBUG nova.network.neutron [-] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.022586] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b873a45b-0bd7-4581-ab30-b849f3b3c86b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Lock "e086d79d-29e6-41bf-a139-0b680ca0021d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.807s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.030592] env[62692]: DEBUG nova.compute.utils [None req-2650f739-3dd3-49fb-ae92-85115406f198 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Can not refresh info_cache because instance was not found {{(pid=62692) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1024}} [ 1071.047372] env[62692]: DEBUG nova.network.neutron [None req-2650f739-3dd3-49fb-ae92-85115406f198 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1071.056023] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-80d28f16-0cbc-4685-b669-ff5babbde8e9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.065227] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2fbccc7-44de-4b8e-92e9-9675931a0d73 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.096305] env[62692]: DEBUG nova.compute.manager [req-f5a02caa-43a1-4253-a795-f6399a5f8cdc req-fff02e25-7061-42b9-a982-480a6cec594b service nova] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Detach interface failed, port_id=898dbc93-ff46-4a9e-9214-060a76c66507, reason: Instance ca56c2ec-95fc-44fc-856f-e395bda1703c could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1071.112409] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.129236] env[62692]: DEBUG nova.network.neutron [None req-2650f739-3dd3-49fb-ae92-85115406f198 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.193013] env[62692]: DEBUG oslo_vmware.api [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141740, 'name': PowerOnVM_Task, 'duration_secs': 0.802194} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.193306] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1071.193516] env[62692]: INFO nova.compute.manager [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Took 9.97 seconds to spawn the instance on the hypervisor. [ 1071.193698] env[62692]: DEBUG nova.compute.manager [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1071.194482] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e123cfc-804f-45b6-99bd-1d995caf4f3e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.280368] env[62692]: INFO nova.compute.manager [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Rebuilding instance [ 1071.304574] env[62692]: DEBUG nova.scheduler.client.report [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1071.326777] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "refresh_cache-a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.326777] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquired lock "refresh_cache-a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.326777] env[62692]: DEBUG nova.network.neutron [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1071.328739] env[62692]: DEBUG nova.compute.manager [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1071.330114] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d424211-5f0f-458c-9778-e069140c6184 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.373389] env[62692]: DEBUG oslo_vmware.api [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141739, 'name': ReconfigVM_Task, 'duration_secs': 1.189315} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.373749] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Reconfigured VM instance instance-0000005a to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1071.375117] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9021491-e2b6-4658-8643-3bed38bab7f1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.399787] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da/9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1071.400294] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0862c89a-6b9c-40f9-9825-0b8a5128e759 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.419359] env[62692]: DEBUG oslo_vmware.api [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1071.419359] env[62692]: value = "task-1141741" [ 1071.419359] env[62692]: _type = "Task" [ 1071.419359] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.427274] env[62692]: DEBUG oslo_vmware.api [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141741, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.442324] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1071.442430] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Starting heal instance info cache {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1071.464675] env[62692]: INFO nova.compute.manager [-] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Took 1.22 seconds to deallocate network for instance. [ 1071.631573] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2650f739-3dd3-49fb-ae92-85115406f198 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Releasing lock "refresh_cache-af260f9a-4e9b-4af1-90da-c619c4755eca" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.632205] env[62692]: DEBUG nova.compute.manager [None req-2650f739-3dd3-49fb-ae92-85115406f198 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1071.632539] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2650f739-3dd3-49fb-ae92-85115406f198 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1071.632920] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-795c1058-e873-48b7-a16c-ce5f96bf4e72 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.642727] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dbaa5db-3485-487a-88db-86f9de923f51 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.674930] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-2650f739-3dd3-49fb-ae92-85115406f198 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance af260f9a-4e9b-4af1-90da-c619c4755eca could not be found. [ 1071.675318] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2650f739-3dd3-49fb-ae92-85115406f198 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1071.675636] env[62692]: INFO nova.compute.manager [None req-2650f739-3dd3-49fb-ae92-85115406f198 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Took 0.04 seconds to destroy the instance on the hypervisor. [ 1071.676056] env[62692]: DEBUG oslo.service.loopingcall [None req-2650f739-3dd3-49fb-ae92-85115406f198 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1071.676417] env[62692]: DEBUG nova.compute.manager [-] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1071.676585] env[62692]: DEBUG nova.network.neutron [-] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1071.692767] env[62692]: DEBUG nova.network.neutron [-] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1071.709656] env[62692]: INFO nova.compute.manager [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Took 31.67 seconds to build instance. [ 1071.813214] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.840s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.816476] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.897s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.818736] env[62692]: INFO nova.compute.claims [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1071.832939] env[62692]: INFO nova.scheduler.client.report [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Deleted allocations for instance a1951dba-ec5a-4849-9e22-37fba08ec60c [ 1071.844166] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1071.844166] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b1caba80-77e1-4ed6-8ccf-d01a59460674 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.851455] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Waiting for the task: (returnval){ [ 1071.851455] env[62692]: value = "task-1141742" [ 1071.851455] env[62692]: _type = "Task" [ 1071.851455] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.860615] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141742, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.868191] env[62692]: DEBUG oslo_concurrency.lockutils [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquiring lock "1b753882-9118-479c-aa3e-cbb8bdf2e086" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.868530] env[62692]: DEBUG oslo_concurrency.lockutils [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Lock "1b753882-9118-479c-aa3e-cbb8bdf2e086" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.868753] env[62692]: DEBUG oslo_concurrency.lockutils [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquiring lock "1b753882-9118-479c-aa3e-cbb8bdf2e086-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.868988] env[62692]: DEBUG oslo_concurrency.lockutils [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Lock "1b753882-9118-479c-aa3e-cbb8bdf2e086-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.869182] env[62692]: DEBUG oslo_concurrency.lockutils [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Lock "1b753882-9118-479c-aa3e-cbb8bdf2e086-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.871429] env[62692]: INFO nova.compute.manager [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Terminating instance [ 1071.873429] env[62692]: DEBUG nova.compute.manager [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1071.873630] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1071.874479] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19310acd-5f72-4062-a5c1-9b8e363cdfba {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.878042] env[62692]: DEBUG nova.network.neutron [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1071.887016] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1071.887016] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-83d013eb-f053-4af9-97d6-36f6eda23d65 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.894447] env[62692]: DEBUG oslo_vmware.api [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Waiting for the task: (returnval){ [ 1071.894447] env[62692]: value = "task-1141743" [ 1071.894447] env[62692]: _type = "Task" [ 1071.894447] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.904129] env[62692]: DEBUG oslo_vmware.api [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141743, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.930324] env[62692]: DEBUG oslo_vmware.api [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141741, 'name': ReconfigVM_Task, 'duration_secs': 0.271096} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.930698] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Reconfigured VM instance instance-0000005a to attach disk [datastore2] 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da/9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1071.931420] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Updating instance '9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da' progress to 50 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1071.970941] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.086427] env[62692]: DEBUG nova.network.neutron [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Updating instance_info_cache with network_info: [{"id": "c3aed689-318a-45c3-a236-6ca6b1bace9b", "address": "fa:16:3e:23:aa:61", "network": {"id": "e9c538a3-7723-42cb-85b7-5a1597e49cef", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1878627765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71b00c566cea4b21811ea1e1ef5ba0f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc3aed689-31", "ovs_interfaceid": "c3aed689-318a-45c3-a236-6ca6b1bace9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.195592] env[62692]: DEBUG nova.network.neutron [-] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.210872] env[62692]: DEBUG oslo_concurrency.lockutils [None req-edf8c9a3-e7ef-4ff5-8d24-9b68a27c7c66 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "90932f05-fad3-4c6d-87ae-ab059351b0be" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.185s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.343026] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f84e39e7-ddd9-4d05-87cc-7e3ae5ecfcfb tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "a1951dba-ec5a-4849-9e22-37fba08ec60c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.936s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.362417] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141742, 'name': PowerOffVM_Task, 'duration_secs': 0.194617} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.362417] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1072.362770] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1072.363633] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e3576d6-3fe0-4b80-8861-75c4ae961ee9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.370689] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1072.371051] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2e0cf67b-98e0-4b5c-bc72-73401d17240d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.395801] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1072.396214] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1072.396557] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Deleting the datastore file [datastore2] e086d79d-29e6-41bf-a139-0b680ca0021d {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1072.400260] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f2b28e5f-4591-4c73-ab13-4999bcd17c93 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.408039] env[62692]: DEBUG oslo_vmware.api [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141743, 'name': PowerOffVM_Task, 'duration_secs': 0.18145} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.409267] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1072.409558] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1072.409975] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Waiting for the task: (returnval){ [ 1072.409975] env[62692]: value = "task-1141745" [ 1072.409975] env[62692]: _type = "Task" [ 1072.409975] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.412551] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5e1a52af-5d1f-4859-915e-895d66514979 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.419710] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141745, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.440021] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4e05885-0ecb-4132-9e2b-e3979fa799ff {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.458346] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8998d12e-60a1-429a-baad-02e8256e87fd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.477733] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Updating instance '9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da' progress to 67 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1072.482681] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1072.483064] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1072.486274] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Deleting the datastore file [datastore2] 1b753882-9118-479c-aa3e-cbb8bdf2e086 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1072.486274] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7ba91a2d-9baf-4be1-8028-851b423a33d6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.489830] env[62692]: DEBUG oslo_vmware.api [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Waiting for the task: (returnval){ [ 1072.489830] env[62692]: value = "task-1141747" [ 1072.489830] env[62692]: _type = "Task" [ 1072.489830] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.498188] env[62692]: DEBUG oslo_vmware.api [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141747, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.589637] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Releasing lock "refresh_cache-a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.590044] env[62692]: DEBUG nova.compute.manager [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Instance network_info: |[{"id": "c3aed689-318a-45c3-a236-6ca6b1bace9b", "address": "fa:16:3e:23:aa:61", "network": {"id": "e9c538a3-7723-42cb-85b7-5a1597e49cef", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1878627765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71b00c566cea4b21811ea1e1ef5ba0f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc3aed689-31", "ovs_interfaceid": "c3aed689-318a-45c3-a236-6ca6b1bace9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1072.590466] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:23:aa:61', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dd96b39f-bd2e-48d1-85c3-577cf97f08c8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c3aed689-318a-45c3-a236-6ca6b1bace9b', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1072.601414] env[62692]: DEBUG oslo.service.loopingcall [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1072.604547] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1072.604547] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4ec7084f-b6a2-4c47-81b6-17170b1d7c12 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.622367] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1072.622367] env[62692]: value = "task-1141748" [ 1072.622367] env[62692]: _type = "Task" [ 1072.622367] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.634751] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141748, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.643045] env[62692]: DEBUG nova.compute.manager [req-dbb5331d-7178-4d5c-87cf-105cf24e38e6 req-9aa8b293-fef6-4f19-a894-989894f91e9f service nova] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Received event network-changed-21e76034-6c0f-4938-afda-0d005cf45759 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1072.643162] env[62692]: DEBUG nova.compute.manager [req-dbb5331d-7178-4d5c-87cf-105cf24e38e6 req-9aa8b293-fef6-4f19-a894-989894f91e9f service nova] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Refreshing instance network info cache due to event network-changed-21e76034-6c0f-4938-afda-0d005cf45759. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1072.643449] env[62692]: DEBUG oslo_concurrency.lockutils [req-dbb5331d-7178-4d5c-87cf-105cf24e38e6 req-9aa8b293-fef6-4f19-a894-989894f91e9f service nova] Acquiring lock "refresh_cache-90932f05-fad3-4c6d-87ae-ab059351b0be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.643487] env[62692]: DEBUG oslo_concurrency.lockutils [req-dbb5331d-7178-4d5c-87cf-105cf24e38e6 req-9aa8b293-fef6-4f19-a894-989894f91e9f service nova] Acquired lock "refresh_cache-90932f05-fad3-4c6d-87ae-ab059351b0be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.643652] env[62692]: DEBUG nova.network.neutron [req-dbb5331d-7178-4d5c-87cf-105cf24e38e6 req-9aa8b293-fef6-4f19-a894-989894f91e9f service nova] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Refreshing network info cache for port 21e76034-6c0f-4938-afda-0d005cf45759 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1072.698957] env[62692]: INFO nova.compute.manager [-] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Took 1.02 seconds to deallocate network for instance. [ 1072.799784] env[62692]: DEBUG nova.compute.manager [req-3a8fbbfa-9dd1-4f4a-bd9f-dd5e3cca1fe4 req-3c984725-2cfb-475a-8d4f-e34f6b12646a service nova] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Received event network-changed-c3aed689-318a-45c3-a236-6ca6b1bace9b {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1072.800078] env[62692]: DEBUG nova.compute.manager [req-3a8fbbfa-9dd1-4f4a-bd9f-dd5e3cca1fe4 req-3c984725-2cfb-475a-8d4f-e34f6b12646a service nova] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Refreshing instance network info cache due to event network-changed-c3aed689-318a-45c3-a236-6ca6b1bace9b. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1072.800170] env[62692]: DEBUG oslo_concurrency.lockutils [req-3a8fbbfa-9dd1-4f4a-bd9f-dd5e3cca1fe4 req-3c984725-2cfb-475a-8d4f-e34f6b12646a service nova] Acquiring lock "refresh_cache-a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.800343] env[62692]: DEBUG oslo_concurrency.lockutils [req-3a8fbbfa-9dd1-4f4a-bd9f-dd5e3cca1fe4 req-3c984725-2cfb-475a-8d4f-e34f6b12646a service nova] Acquired lock "refresh_cache-a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.800579] env[62692]: DEBUG nova.network.neutron [req-3a8fbbfa-9dd1-4f4a-bd9f-dd5e3cca1fe4 req-3c984725-2cfb-475a-8d4f-e34f6b12646a service nova] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Refreshing network info cache for port c3aed689-318a-45c3-a236-6ca6b1bace9b {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1072.923139] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141745, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.188428} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.923429] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1072.923926] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1072.923926] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1073.001460] env[62692]: DEBUG oslo_vmware.api [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Task: {'id': task-1141747, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.210623} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.004059] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1073.004268] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1073.004455] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1073.004634] env[62692]: INFO nova.compute.manager [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1073.004885] env[62692]: DEBUG oslo.service.loopingcall [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1073.005461] env[62692]: DEBUG nova.compute.manager [-] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1073.005568] env[62692]: DEBUG nova.network.neutron [-] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1073.022537] env[62692]: DEBUG nova.network.neutron [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Port d5ff0422-e9b3-455c-bd28-422096794288 binding to destination host cpu-1 is already ACTIVE {{(pid=62692) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1073.043558] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c189566-b059-4626-8142-d03b11fdf8ed {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.051991] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27017423-f57a-432f-bd4b-dc73dffd32c8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.085762] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eafb235f-fffa-441b-9a00-fc8df604d0c1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.093252] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-864a47b8-9f93-4e04-826e-bf8d261954fe {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.106713] env[62692]: DEBUG nova.compute.provider_tree [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1073.135579] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141748, 'name': CreateVM_Task, 'duration_secs': 0.345995} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.135579] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1073.135579] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.135579] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.135579] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1073.135579] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba117a98-5fe5-40fa-80b2-99e3873bde13 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.140872] env[62692]: DEBUG oslo_vmware.api [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1073.140872] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5210d6df-f53b-ab35-35c0-d017358c461e" [ 1073.140872] env[62692]: _type = "Task" [ 1073.140872] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.151712] env[62692]: DEBUG oslo_vmware.api [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5210d6df-f53b-ab35-35c0-d017358c461e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.206689] env[62692]: INFO nova.compute.manager [None req-2650f739-3dd3-49fb-ae92-85115406f198 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Instance disappeared during terminate [ 1073.208068] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2650f739-3dd3-49fb-ae92-85115406f198 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Lock "af260f9a-4e9b-4af1-90da-c619c4755eca" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.690s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.480648] env[62692]: DEBUG nova.network.neutron [req-dbb5331d-7178-4d5c-87cf-105cf24e38e6 req-9aa8b293-fef6-4f19-a894-989894f91e9f service nova] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Updated VIF entry in instance network info cache for port 21e76034-6c0f-4938-afda-0d005cf45759. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1073.481036] env[62692]: DEBUG nova.network.neutron [req-dbb5331d-7178-4d5c-87cf-105cf24e38e6 req-9aa8b293-fef6-4f19-a894-989894f91e9f service nova] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Updating instance_info_cache with network_info: [{"id": "21e76034-6c0f-4938-afda-0d005cf45759", "address": "fa:16:3e:2d:c0:4a", "network": {"id": "77695f27-3027-4a92-8b9b-0a90cdc53f62", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-337681783-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697b4281872f4547ad3d051e644b95ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21e76034-6c", "ovs_interfaceid": "21e76034-6c0f-4938-afda-0d005cf45759", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.573157] env[62692]: DEBUG nova.network.neutron [req-3a8fbbfa-9dd1-4f4a-bd9f-dd5e3cca1fe4 req-3c984725-2cfb-475a-8d4f-e34f6b12646a service nova] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Updated VIF entry in instance network info cache for port c3aed689-318a-45c3-a236-6ca6b1bace9b. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1073.573641] env[62692]: DEBUG nova.network.neutron [req-3a8fbbfa-9dd1-4f4a-bd9f-dd5e3cca1fe4 req-3c984725-2cfb-475a-8d4f-e34f6b12646a service nova] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Updating instance_info_cache with network_info: [{"id": "c3aed689-318a-45c3-a236-6ca6b1bace9b", "address": "fa:16:3e:23:aa:61", "network": {"id": "e9c538a3-7723-42cb-85b7-5a1597e49cef", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1878627765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71b00c566cea4b21811ea1e1ef5ba0f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc3aed689-31", "ovs_interfaceid": "c3aed689-318a-45c3-a236-6ca6b1bace9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.602446] env[62692]: DEBUG oslo_concurrency.lockutils [None req-efac3b0a-0872-4fcd-b4da-80753318f032 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "e8f645da-3b4a-4f21-9a17-aebdc8d112df" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.602653] env[62692]: DEBUG oslo_concurrency.lockutils [None req-efac3b0a-0872-4fcd-b4da-80753318f032 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "e8f645da-3b4a-4f21-9a17-aebdc8d112df" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.609777] env[62692]: DEBUG nova.scheduler.client.report [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1073.652014] env[62692]: DEBUG oslo_vmware.api [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5210d6df-f53b-ab35-35c0-d017358c461e, 'name': SearchDatastore_Task, 'duration_secs': 0.033269} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.653026] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.653026] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1073.653026] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.653026] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.653226] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1073.653442] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-70426977-93ec-4018-ac7b-a053899c2a31 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.661556] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1073.661749] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1073.662504] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4d85eb1-52f6-4180-a2eb-2ddc0252e18f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.669081] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "f92fae48-7285-4b69-97c9-2360227c18fc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.669320] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "f92fae48-7285-4b69-97c9-2360227c18fc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.672915] env[62692]: DEBUG oslo_vmware.api [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1073.672915] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52b48c00-5474-aca7-ba43-d27918a201be" [ 1073.672915] env[62692]: _type = "Task" [ 1073.672915] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.684315] env[62692]: DEBUG oslo_vmware.api [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52b48c00-5474-aca7-ba43-d27918a201be, 'name': SearchDatastore_Task, 'duration_secs': 0.009076} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.685131] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc243df6-a8ed-4bcb-b0fe-98783dcb7f3b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.690735] env[62692]: DEBUG oslo_vmware.api [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1073.690735] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]520904f5-bbcb-1239-7afa-8b6037bdb1a5" [ 1073.690735] env[62692]: _type = "Task" [ 1073.690735] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.698995] env[62692]: DEBUG oslo_vmware.api [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]520904f5-bbcb-1239-7afa-8b6037bdb1a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.792056] env[62692]: DEBUG nova.network.neutron [-] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.968805] env[62692]: DEBUG nova.virt.hardware [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1073.969094] env[62692]: DEBUG nova.virt.hardware [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1073.969267] env[62692]: DEBUG nova.virt.hardware [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1073.969459] env[62692]: DEBUG nova.virt.hardware [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1073.969614] env[62692]: DEBUG nova.virt.hardware [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1073.969768] env[62692]: DEBUG nova.virt.hardware [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1073.969982] env[62692]: DEBUG nova.virt.hardware [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1073.970167] env[62692]: DEBUG nova.virt.hardware [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1073.970342] env[62692]: DEBUG nova.virt.hardware [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1073.970541] env[62692]: DEBUG nova.virt.hardware [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1073.970729] env[62692]: DEBUG nova.virt.hardware [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1073.972998] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e20ef51d-aa87-486c-9ba6-34cf6776535d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.981755] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f332d52-0ddf-4929-8e69-0e8444824e7e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.986855] env[62692]: DEBUG oslo_concurrency.lockutils [req-dbb5331d-7178-4d5c-87cf-105cf24e38e6 req-9aa8b293-fef6-4f19-a894-989894f91e9f service nova] Releasing lock "refresh_cache-90932f05-fad3-4c6d-87ae-ab059351b0be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.997919] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Instance VIF info [] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1074.003627] env[62692]: DEBUG oslo.service.loopingcall [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1074.004120] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1074.004355] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-64d633db-c73c-432c-83c9-d44f1a8d2163 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.022215] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1074.022215] env[62692]: value = "task-1141749" [ 1074.022215] env[62692]: _type = "Task" [ 1074.022215] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.039326] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141749, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.046040] env[62692]: DEBUG oslo_concurrency.lockutils [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.046190] env[62692]: DEBUG oslo_concurrency.lockutils [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.046365] env[62692]: DEBUG oslo_concurrency.lockutils [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.077460] env[62692]: DEBUG oslo_concurrency.lockutils [req-3a8fbbfa-9dd1-4f4a-bd9f-dd5e3cca1fe4 req-3c984725-2cfb-475a-8d4f-e34f6b12646a service nova] Releasing lock "refresh_cache-a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.105685] env[62692]: DEBUG nova.compute.utils [None req-efac3b0a-0872-4fcd-b4da-80753318f032 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1074.116384] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.300s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.116969] env[62692]: DEBUG nova.compute.manager [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1074.120066] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.797s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.121919] env[62692]: INFO nova.compute.claims [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1074.174109] env[62692]: DEBUG nova.compute.manager [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1074.202020] env[62692]: DEBUG oslo_vmware.api [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]520904f5-bbcb-1239-7afa-8b6037bdb1a5, 'name': SearchDatastore_Task, 'duration_secs': 0.008868} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.202217] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.202494] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970/a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1074.202787] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4e24c548-7cce-4835-9305-5c955493aa44 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.211565] env[62692]: DEBUG oslo_vmware.api [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1074.211565] env[62692]: value = "task-1141750" [ 1074.211565] env[62692]: _type = "Task" [ 1074.211565] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.220558] env[62692]: DEBUG oslo_vmware.api [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141750, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.293916] env[62692]: INFO nova.compute.manager [-] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Took 1.29 seconds to deallocate network for instance. [ 1074.533805] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141749, 'name': CreateVM_Task, 'duration_secs': 0.367104} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.534028] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1074.534554] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1074.534778] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1074.535220] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1074.535537] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-213d5c57-4df3-4162-b81f-5865a30ed6cd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.541071] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Waiting for the task: (returnval){ [ 1074.541071] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52d3a517-6950-f7b7-8f1c-10defd6d4b45" [ 1074.541071] env[62692]: _type = "Task" [ 1074.541071] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.550640] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52d3a517-6950-f7b7-8f1c-10defd6d4b45, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.609234] env[62692]: DEBUG oslo_concurrency.lockutils [None req-efac3b0a-0872-4fcd-b4da-80753318f032 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "e8f645da-3b4a-4f21-9a17-aebdc8d112df" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.627235] env[62692]: DEBUG nova.compute.utils [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1074.630911] env[62692]: DEBUG nova.compute.manager [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1074.631132] env[62692]: DEBUG nova.network.neutron [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1074.676135] env[62692]: DEBUG nova.compute.manager [req-8ceb526b-923c-41d3-8fa9-e89aa36f3d8f req-55e450bf-8411-4ce3-b150-b0c97613760e service nova] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Received event network-vif-deleted-c43f9978-4cfd-467d-86e0-4c8acfefc821 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1074.696481] env[62692]: DEBUG nova.policy [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a251ef5f6cb143178928babc75a0fef0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ba97a775e99a463c845518b0df8cd3e2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1074.700378] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.725390] env[62692]: DEBUG oslo_vmware.api [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141750, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.801174] env[62692]: DEBUG oslo_concurrency.lockutils [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.026613] env[62692]: DEBUG nova.network.neutron [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Successfully created port: d51ffd3f-c023-45e2-8bfa-1424b7ccb188 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1075.053930] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52d3a517-6950-f7b7-8f1c-10defd6d4b45, 'name': SearchDatastore_Task, 'duration_secs': 0.04871} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.054160] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1075.054394] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1075.054629] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1075.054781] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.054969] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1075.056697] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b632ab08-ffb9-440d-8027-fe5935ee63c1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.065873] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1075.066075] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1075.066782] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8ca3c3a-ca93-4332-99e0-63f58094b077 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.072321] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Waiting for the task: (returnval){ [ 1075.072321] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52efe746-a239-c4ed-228a-2fc7842fda62" [ 1075.072321] env[62692]: _type = "Task" [ 1075.072321] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.080391] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52efe746-a239-c4ed-228a-2fc7842fda62, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.086675] env[62692]: DEBUG oslo_concurrency.lockutils [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "refresh_cache-9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1075.086871] env[62692]: DEBUG oslo_concurrency.lockutils [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired lock "refresh_cache-9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.087074] env[62692]: DEBUG nova.network.neutron [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1075.132094] env[62692]: DEBUG nova.compute.manager [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1075.230905] env[62692]: DEBUG oslo_vmware.api [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141750, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.654329} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.231234] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970/a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1075.231462] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1075.231725] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-75e990f1-e307-4e2c-a041-8bbafe2eb4c2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.242361] env[62692]: DEBUG oslo_vmware.api [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1075.242361] env[62692]: value = "task-1141751" [ 1075.242361] env[62692]: _type = "Task" [ 1075.242361] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.247469] env[62692]: DEBUG oslo_vmware.api [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141751, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.359238] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8291e518-811d-4164-a8f8-162940638acd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.367745] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5487f1aa-9453-4cf4-87f3-eb072e2a63e9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.401842] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59052f2e-2244-4c2d-945e-3afacb9379f4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.409558] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-052d2bfa-2f55-4680-ad7c-230a6fabaf1d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.423099] env[62692]: DEBUG nova.compute.provider_tree [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1075.583547] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52efe746-a239-c4ed-228a-2fc7842fda62, 'name': SearchDatastore_Task, 'duration_secs': 0.022158} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.584020] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a5ced31-9570-448b-ac4b-15bac127b61c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.589097] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Waiting for the task: (returnval){ [ 1075.589097] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52008843-6743-ed99-32bb-51c711689cf1" [ 1075.589097] env[62692]: _type = "Task" [ 1075.589097] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.601272] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52008843-6743-ed99-32bb-51c711689cf1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.681491] env[62692]: DEBUG oslo_concurrency.lockutils [None req-efac3b0a-0872-4fcd-b4da-80753318f032 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "e8f645da-3b4a-4f21-9a17-aebdc8d112df" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.681755] env[62692]: DEBUG oslo_concurrency.lockutils [None req-efac3b0a-0872-4fcd-b4da-80753318f032 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "e8f645da-3b4a-4f21-9a17-aebdc8d112df" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.681987] env[62692]: INFO nova.compute.manager [None req-efac3b0a-0872-4fcd-b4da-80753318f032 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Attaching volume ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60 to /dev/sdb [ 1075.712392] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4cb57ef-7bbb-4527-92d0-f86ab3ae22d8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.719332] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0e69b21-3702-4b51-be7d-0f0303be5e91 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.733504] env[62692]: DEBUG nova.virt.block_device [None req-efac3b0a-0872-4fcd-b4da-80753318f032 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Updating existing volume attachment record: 92ef4bfe-08c4-41c1-9422-22d666392f65 {{(pid=62692) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1075.747991] env[62692]: DEBUG oslo_vmware.api [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141751, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062579} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.748291] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1075.749117] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02adb488-bef8-4169-bd8b-3e882951cd56 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.771299] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970/a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1075.773949] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-60756af8-aa8b-4cb1-8856-1f7040650876 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.800662] env[62692]: DEBUG oslo_vmware.api [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1075.800662] env[62692]: value = "task-1141752" [ 1075.800662] env[62692]: _type = "Task" [ 1075.800662] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.808758] env[62692]: DEBUG oslo_vmware.api [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141752, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.827790] env[62692]: DEBUG nova.network.neutron [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Updating instance_info_cache with network_info: [{"id": "d5ff0422-e9b3-455c-bd28-422096794288", "address": "fa:16:3e:46:38:ff", "network": {"id": "f6352999-f943-431d-b54a-c9dc1d810a85", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-121037111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9cb8ae08824245a7911de7455e7fcb68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c842425c-544e-4ce2-9657-512723bd318e", "external-id": "nsx-vlan-transportzone-80", "segmentation_id": 80, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5ff0422-e9", "ovs_interfaceid": "d5ff0422-e9b3-455c-bd28-422096794288", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.926424] env[62692]: DEBUG nova.scheduler.client.report [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1076.100629] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52008843-6743-ed99-32bb-51c711689cf1, 'name': SearchDatastore_Task, 'duration_secs': 0.065128} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.100977] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1076.101211] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] e086d79d-29e6-41bf-a139-0b680ca0021d/e086d79d-29e6-41bf-a139-0b680ca0021d.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1076.101483] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a22e2d7b-5853-4ce8-b817-ffb1cd31d331 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.107154] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Waiting for the task: (returnval){ [ 1076.107154] env[62692]: value = "task-1141756" [ 1076.107154] env[62692]: _type = "Task" [ 1076.107154] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.114716] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141756, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.145020] env[62692]: DEBUG nova.compute.manager [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1076.173683] env[62692]: DEBUG nova.virt.hardware [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1076.173964] env[62692]: DEBUG nova.virt.hardware [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1076.174108] env[62692]: DEBUG nova.virt.hardware [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1076.174315] env[62692]: DEBUG nova.virt.hardware [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1076.174492] env[62692]: DEBUG nova.virt.hardware [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1076.174672] env[62692]: DEBUG nova.virt.hardware [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1076.174886] env[62692]: DEBUG nova.virt.hardware [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1076.175078] env[62692]: DEBUG nova.virt.hardware [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1076.175280] env[62692]: DEBUG nova.virt.hardware [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1076.175467] env[62692]: DEBUG nova.virt.hardware [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1076.175660] env[62692]: DEBUG nova.virt.hardware [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1076.176551] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83e399da-c343-4d51-943d-89afaab3d15c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.184418] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9e0c7ad-780d-4859-bdd7-e3b5589798e0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.313343] env[62692]: DEBUG oslo_vmware.api [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141752, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.331841] env[62692]: DEBUG oslo_concurrency.lockutils [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Releasing lock "refresh_cache-9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1076.432155] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.312s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.433112] env[62692]: DEBUG nova.compute.manager [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1076.436878] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.325s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.439020] env[62692]: INFO nova.compute.claims [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1076.626430] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141756, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.704100] env[62692]: DEBUG nova.compute.manager [req-16462391-d85c-43a9-9979-23f0d46dd727 req-94fcdcab-6cf5-4702-86a1-23c85cc44682 service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Received event network-vif-plugged-d51ffd3f-c023-45e2-8bfa-1424b7ccb188 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1076.704100] env[62692]: DEBUG oslo_concurrency.lockutils [req-16462391-d85c-43a9-9979-23f0d46dd727 req-94fcdcab-6cf5-4702-86a1-23c85cc44682 service nova] Acquiring lock "94a07505-661f-4595-9024-8a9da9f2c2fc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.704100] env[62692]: DEBUG oslo_concurrency.lockutils [req-16462391-d85c-43a9-9979-23f0d46dd727 req-94fcdcab-6cf5-4702-86a1-23c85cc44682 service nova] Lock "94a07505-661f-4595-9024-8a9da9f2c2fc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.704100] env[62692]: DEBUG oslo_concurrency.lockutils [req-16462391-d85c-43a9-9979-23f0d46dd727 req-94fcdcab-6cf5-4702-86a1-23c85cc44682 service nova] Lock "94a07505-661f-4595-9024-8a9da9f2c2fc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.704100] env[62692]: DEBUG nova.compute.manager [req-16462391-d85c-43a9-9979-23f0d46dd727 req-94fcdcab-6cf5-4702-86a1-23c85cc44682 service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] No waiting events found dispatching network-vif-plugged-d51ffd3f-c023-45e2-8bfa-1424b7ccb188 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1076.704418] env[62692]: WARNING nova.compute.manager [req-16462391-d85c-43a9-9979-23f0d46dd727 req-94fcdcab-6cf5-4702-86a1-23c85cc44682 service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Received unexpected event network-vif-plugged-d51ffd3f-c023-45e2-8bfa-1424b7ccb188 for instance with vm_state building and task_state spawning. [ 1076.732046] env[62692]: DEBUG nova.network.neutron [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Successfully updated port: d51ffd3f-c023-45e2-8bfa-1424b7ccb188 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1076.811953] env[62692]: DEBUG oslo_vmware.api [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141752, 'name': ReconfigVM_Task, 'duration_secs': 0.548319} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.812074] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Reconfigured VM instance instance-00000060 to attach disk [datastore2] a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970/a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1076.813394] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-875a7e52-d517-434d-8dec-ead71ec46611 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.820060] env[62692]: DEBUG oslo_vmware.api [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1076.820060] env[62692]: value = "task-1141757" [ 1076.820060] env[62692]: _type = "Task" [ 1076.820060] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.827418] env[62692]: DEBUG oslo_vmware.api [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141757, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.857896] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87f88f56-f46a-4f73-86bc-5a80b681b484 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.877876] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de7304e-ba1a-4bf7-9865-48d3f5c0634a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.886365] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Updating instance '9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da' progress to 83 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1076.944388] env[62692]: DEBUG nova.compute.utils [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1076.946351] env[62692]: DEBUG nova.compute.manager [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1076.946536] env[62692]: DEBUG nova.network.neutron [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1076.992877] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Didn't find any instances for network info cache update. {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1076.993243] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.993480] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.993690] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.994217] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.994471] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.994693] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.995067] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62692) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1076.995133] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager.update_available_resource {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.997524] env[62692]: DEBUG nova.policy [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6e9cd6d33e744da4a2f678226b684845', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1037e5d436bd429391159b87a712f6e1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1077.120921] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141756, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.646258} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.121272] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] e086d79d-29e6-41bf-a139-0b680ca0021d/e086d79d-29e6-41bf-a139-0b680ca0021d.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1077.121427] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1077.121702] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9ed93866-94d3-445d-8d61-8e2085bb54cd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.127322] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Waiting for the task: (returnval){ [ 1077.127322] env[62692]: value = "task-1141758" [ 1077.127322] env[62692]: _type = "Task" [ 1077.127322] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.136962] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141758, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.236304] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "refresh_cache-94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.236304] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "refresh_cache-94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.236304] env[62692]: DEBUG nova.network.neutron [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1077.284749] env[62692]: DEBUG nova.network.neutron [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Successfully created port: a3885c51-6f7b-41e8-8948-41fac442a288 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1077.330201] env[62692]: DEBUG oslo_vmware.api [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141757, 'name': Rename_Task, 'duration_secs': 0.17287} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.330526] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1077.330778] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2580187f-0b1e-4e9e-b9d9-18feac1a58f2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.337557] env[62692]: DEBUG oslo_vmware.api [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1077.337557] env[62692]: value = "task-1141759" [ 1077.337557] env[62692]: _type = "Task" [ 1077.337557] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.344845] env[62692]: DEBUG oslo_vmware.api [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141759, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.393522] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1077.393846] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-72428f52-9fe8-4ca3-9a00-cf4c23f0223a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.400716] env[62692]: DEBUG oslo_vmware.api [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1077.400716] env[62692]: value = "task-1141760" [ 1077.400716] env[62692]: _type = "Task" [ 1077.400716] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.412036] env[62692]: DEBUG oslo_vmware.api [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141760, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.450084] env[62692]: DEBUG nova.compute.manager [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1077.501600] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.637060] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141758, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066276} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.637399] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1077.638120] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-119a98a9-c3f1-4309-85a0-4d364db25055 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.643419] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48ffab94-25f7-4358-bb36-e6f266f65c70 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.660942] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] e086d79d-29e6-41bf-a139-0b680ca0021d/e086d79d-29e6-41bf-a139-0b680ca0021d.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1077.661757] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b725898b-f57d-414e-bf0a-5abb5c8a79f1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.679821] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a03d92c-e9ed-4a38-abc7-241c38eb7b6d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.684402] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Waiting for the task: (returnval){ [ 1077.684402] env[62692]: value = "task-1141761" [ 1077.684402] env[62692]: _type = "Task" [ 1077.684402] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.718983] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd8e037a-5919-49f8-8205-19c0d4777b21 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.727340] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141761, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.733960] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c3a04cb-16a5-4a74-a7ea-2da8855eb211 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.755109] env[62692]: DEBUG nova.compute.provider_tree [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1077.777231] env[62692]: DEBUG nova.network.neutron [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1077.847729] env[62692]: DEBUG oslo_vmware.api [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141759, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.911143] env[62692]: DEBUG oslo_vmware.api [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141760, 'name': PowerOnVM_Task, 'duration_secs': 0.422355} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.911564] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1077.911784] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-53600d39-bffd-4bf8-971d-a92af9dc899d tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Updating instance '9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da' progress to 100 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1077.993045] env[62692]: DEBUG nova.network.neutron [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Updating instance_info_cache with network_info: [{"id": "d51ffd3f-c023-45e2-8bfa-1424b7ccb188", "address": "fa:16:3e:a8:44:f1", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd51ffd3f-c0", "ovs_interfaceid": "d51ffd3f-c023-45e2-8bfa-1424b7ccb188", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.195780] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141761, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.258390] env[62692]: DEBUG nova.scheduler.client.report [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1078.347991] env[62692]: DEBUG oslo_vmware.api [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141759, 'name': PowerOnVM_Task, 'duration_secs': 0.577633} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.348288] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1078.348506] env[62692]: INFO nova.compute.manager [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Took 8.29 seconds to spawn the instance on the hypervisor. [ 1078.348721] env[62692]: DEBUG nova.compute.manager [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1078.349576] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5d87606-402e-4002-85eb-bb6ff53f8214 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.462116] env[62692]: DEBUG nova.compute.manager [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1078.488842] env[62692]: DEBUG nova.virt.hardware [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1078.489097] env[62692]: DEBUG nova.virt.hardware [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1078.489268] env[62692]: DEBUG nova.virt.hardware [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1078.489458] env[62692]: DEBUG nova.virt.hardware [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1078.489612] env[62692]: DEBUG nova.virt.hardware [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1078.489781] env[62692]: DEBUG nova.virt.hardware [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1078.489999] env[62692]: DEBUG nova.virt.hardware [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1078.490181] env[62692]: DEBUG nova.virt.hardware [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1078.490357] env[62692]: DEBUG nova.virt.hardware [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1078.490548] env[62692]: DEBUG nova.virt.hardware [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1078.490744] env[62692]: DEBUG nova.virt.hardware [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1078.491640] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33f1892e-b872-45ae-8d03-707c9dab0905 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.494801] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "refresh_cache-94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.495069] env[62692]: DEBUG nova.compute.manager [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Instance network_info: |[{"id": "d51ffd3f-c023-45e2-8bfa-1424b7ccb188", "address": "fa:16:3e:a8:44:f1", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd51ffd3f-c0", "ovs_interfaceid": "d51ffd3f-c023-45e2-8bfa-1424b7ccb188", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1078.495451] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a8:44:f1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6a31b224-77d7-48a4-af87-312758611b8e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd51ffd3f-c023-45e2-8bfa-1424b7ccb188', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1078.503758] env[62692]: DEBUG oslo.service.loopingcall [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1078.506219] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1078.506480] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-337a7c15-4f29-40b6-b264-a10154a34ad5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.521726] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-908b012e-7cb6-4b9c-82eb-4d0b7c4e29df {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.529353] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1078.529353] env[62692]: value = "task-1141763" [ 1078.529353] env[62692]: _type = "Task" [ 1078.529353] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.545560] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141763, 'name': CreateVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.695768] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141761, 'name': ReconfigVM_Task, 'duration_secs': 0.855439} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.696431] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Reconfigured VM instance instance-0000005f to attach disk [datastore2] e086d79d-29e6-41bf-a139-0b680ca0021d/e086d79d-29e6-41bf-a139-0b680ca0021d.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1078.697149] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d48db050-2579-4b32-9372-aaa08ca48865 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.704791] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Waiting for the task: (returnval){ [ 1078.704791] env[62692]: value = "task-1141764" [ 1078.704791] env[62692]: _type = "Task" [ 1078.704791] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.712756] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141764, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.764144] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.327s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.765094] env[62692]: DEBUG nova.compute.manager [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1078.768067] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.797s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.768523] env[62692]: DEBUG nova.objects.instance [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lazy-loading 'resources' on Instance uuid ca56c2ec-95fc-44fc-856f-e395bda1703c {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1078.779937] env[62692]: DEBUG nova.compute.manager [req-62ff8685-9b2e-4b2c-916e-85a84affad0d req-66d43c10-738d-4cab-bbe9-6689032296eb service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Received event network-changed-d51ffd3f-c023-45e2-8bfa-1424b7ccb188 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1078.785364] env[62692]: DEBUG nova.compute.manager [req-62ff8685-9b2e-4b2c-916e-85a84affad0d req-66d43c10-738d-4cab-bbe9-6689032296eb service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Refreshing instance network info cache due to event network-changed-d51ffd3f-c023-45e2-8bfa-1424b7ccb188. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1078.785745] env[62692]: DEBUG oslo_concurrency.lockutils [req-62ff8685-9b2e-4b2c-916e-85a84affad0d req-66d43c10-738d-4cab-bbe9-6689032296eb service nova] Acquiring lock "refresh_cache-94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1078.785987] env[62692]: DEBUG oslo_concurrency.lockutils [req-62ff8685-9b2e-4b2c-916e-85a84affad0d req-66d43c10-738d-4cab-bbe9-6689032296eb service nova] Acquired lock "refresh_cache-94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.786267] env[62692]: DEBUG nova.network.neutron [req-62ff8685-9b2e-4b2c-916e-85a84affad0d req-66d43c10-738d-4cab-bbe9-6689032296eb service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Refreshing network info cache for port d51ffd3f-c023-45e2-8bfa-1424b7ccb188 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1078.867239] env[62692]: INFO nova.compute.manager [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Took 20.42 seconds to build instance. [ 1078.899275] env[62692]: DEBUG nova.network.neutron [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Successfully updated port: a3885c51-6f7b-41e8-8948-41fac442a288 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1079.047628] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141763, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.214842] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141764, 'name': Rename_Task, 'duration_secs': 0.191518} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.215192] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1079.215382] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3616bced-fc9f-41fe-b00e-3104132b66c8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.222285] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Waiting for the task: (returnval){ [ 1079.222285] env[62692]: value = "task-1141765" [ 1079.222285] env[62692]: _type = "Task" [ 1079.222285] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.230182] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141765, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.276023] env[62692]: DEBUG nova.compute.utils [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1079.280951] env[62692]: DEBUG nova.compute.manager [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1079.281124] env[62692]: DEBUG nova.network.neutron [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1079.320374] env[62692]: DEBUG nova.policy [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd0229d6805ee40bd9444c2cba44a6b03', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '43796ce588804516bdbbba3ed5b0e25f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1079.371494] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c9e06793-000c-4aca-af25-5621ebbef4ec tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.936s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.402350] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "refresh_cache-cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1079.402509] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired lock "refresh_cache-cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.402664] env[62692]: DEBUG nova.network.neutron [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1079.481028] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37fc5dc8-0bd1-4cba-94af-a56d1e1d6e31 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.493296] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab38bb8e-92c0-422d-ad60-fe775b538070 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.525591] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48fdb804-df86-463d-be03-73e3dc8730cf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.533391] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e529d2e-0f40-471a-b29d-7fd00fdc5d1f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.551039] env[62692]: DEBUG nova.compute.provider_tree [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1079.557219] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141763, 'name': CreateVM_Task, 'duration_secs': 0.754746} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.557367] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1079.558055] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1079.558280] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.562018] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1079.562018] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfbe4227-18ea-415c-8147-7e220b813a24 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.564259] env[62692]: DEBUG oslo_vmware.api [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1079.564259] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5284fea8-a08e-d514-2bdb-d2a8ed760d5f" [ 1079.564259] env[62692]: _type = "Task" [ 1079.564259] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.572309] env[62692]: DEBUG oslo_vmware.api [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5284fea8-a08e-d514-2bdb-d2a8ed760d5f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.627898] env[62692]: DEBUG nova.network.neutron [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Successfully created port: 06a8d751-3ffa-4fb1-8017-a12d81fc1d98 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1079.663716] env[62692]: DEBUG nova.network.neutron [req-62ff8685-9b2e-4b2c-916e-85a84affad0d req-66d43c10-738d-4cab-bbe9-6689032296eb service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Updated VIF entry in instance network info cache for port d51ffd3f-c023-45e2-8bfa-1424b7ccb188. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1079.664132] env[62692]: DEBUG nova.network.neutron [req-62ff8685-9b2e-4b2c-916e-85a84affad0d req-66d43c10-738d-4cab-bbe9-6689032296eb service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Updating instance_info_cache with network_info: [{"id": "d51ffd3f-c023-45e2-8bfa-1424b7ccb188", "address": "fa:16:3e:a8:44:f1", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd51ffd3f-c0", "ovs_interfaceid": "d51ffd3f-c023-45e2-8bfa-1424b7ccb188", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1079.732462] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141765, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.781643] env[62692]: DEBUG nova.compute.manager [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1079.970471] env[62692]: DEBUG nova.network.neutron [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1080.054791] env[62692]: DEBUG nova.scheduler.client.report [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1080.076607] env[62692]: DEBUG oslo_vmware.api [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5284fea8-a08e-d514-2bdb-d2a8ed760d5f, 'name': SearchDatastore_Task, 'duration_secs': 0.010445} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.080392] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1080.080392] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1080.080605] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1080.080685] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.080891] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1080.082556] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cb51cf98-61fc-40c0-bce8-549863c5f2b7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.091409] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1080.091604] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1080.092656] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c55b2f4d-c3d1-40a5-9777-4841587a3088 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.101874] env[62692]: DEBUG oslo_vmware.api [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1080.101874] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52e52526-3e0f-f9c5-45c1-839afbcc7605" [ 1080.101874] env[62692]: _type = "Task" [ 1080.101874] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.110478] env[62692]: DEBUG oslo_vmware.api [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52e52526-3e0f-f9c5-45c1-839afbcc7605, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.169155] env[62692]: DEBUG oslo_concurrency.lockutils [req-62ff8685-9b2e-4b2c-916e-85a84affad0d req-66d43c10-738d-4cab-bbe9-6689032296eb service nova] Releasing lock "refresh_cache-94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1080.169155] env[62692]: DEBUG nova.compute.manager [req-62ff8685-9b2e-4b2c-916e-85a84affad0d req-66d43c10-738d-4cab-bbe9-6689032296eb service nova] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Received event network-vif-plugged-a3885c51-6f7b-41e8-8948-41fac442a288 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1080.169155] env[62692]: DEBUG oslo_concurrency.lockutils [req-62ff8685-9b2e-4b2c-916e-85a84affad0d req-66d43c10-738d-4cab-bbe9-6689032296eb service nova] Acquiring lock "cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.169155] env[62692]: DEBUG oslo_concurrency.lockutils [req-62ff8685-9b2e-4b2c-916e-85a84affad0d req-66d43c10-738d-4cab-bbe9-6689032296eb service nova] Lock "cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.169155] env[62692]: DEBUG oslo_concurrency.lockutils [req-62ff8685-9b2e-4b2c-916e-85a84affad0d req-66d43c10-738d-4cab-bbe9-6689032296eb service nova] Lock "cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.169357] env[62692]: DEBUG nova.compute.manager [req-62ff8685-9b2e-4b2c-916e-85a84affad0d req-66d43c10-738d-4cab-bbe9-6689032296eb service nova] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] No waiting events found dispatching network-vif-plugged-a3885c51-6f7b-41e8-8948-41fac442a288 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1080.169413] env[62692]: WARNING nova.compute.manager [req-62ff8685-9b2e-4b2c-916e-85a84affad0d req-66d43c10-738d-4cab-bbe9-6689032296eb service nova] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Received unexpected event network-vif-plugged-a3885c51-6f7b-41e8-8948-41fac442a288 for instance with vm_state building and task_state spawning. [ 1080.171206] env[62692]: DEBUG nova.compute.manager [req-a58d989f-ea55-4afc-84f3-755fcbc31b2a req-a8f3581d-445b-4723-8631-b8478b50ab99 service nova] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Received event network-changed-c3aed689-318a-45c3-a236-6ca6b1bace9b {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1080.171826] env[62692]: DEBUG nova.compute.manager [req-a58d989f-ea55-4afc-84f3-755fcbc31b2a req-a8f3581d-445b-4723-8631-b8478b50ab99 service nova] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Refreshing instance network info cache due to event network-changed-c3aed689-318a-45c3-a236-6ca6b1bace9b. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1080.171952] env[62692]: DEBUG oslo_concurrency.lockutils [req-a58d989f-ea55-4afc-84f3-755fcbc31b2a req-a8f3581d-445b-4723-8631-b8478b50ab99 service nova] Acquiring lock "refresh_cache-a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1080.172214] env[62692]: DEBUG oslo_concurrency.lockutils [req-a58d989f-ea55-4afc-84f3-755fcbc31b2a req-a8f3581d-445b-4723-8631-b8478b50ab99 service nova] Acquired lock "refresh_cache-a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.172459] env[62692]: DEBUG nova.network.neutron [req-a58d989f-ea55-4afc-84f3-755fcbc31b2a req-a8f3581d-445b-4723-8631-b8478b50ab99 service nova] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Refreshing network info cache for port c3aed689-318a-45c3-a236-6ca6b1bace9b {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1080.192558] env[62692]: DEBUG nova.network.neutron [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Updating instance_info_cache with network_info: [{"id": "a3885c51-6f7b-41e8-8948-41fac442a288", "address": "fa:16:3e:e4:9b:6f", "network": {"id": "f70727d6-a9aa-44b0-a244-2533b3f96571", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-284105837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1037e5d436bd429391159b87a712f6e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3885c51-6f", "ovs_interfaceid": "a3885c51-6f7b-41e8-8948-41fac442a288", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.236474] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141765, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.284601] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-efac3b0a-0872-4fcd-b4da-80753318f032 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Volume attach. Driver type: vmdk {{(pid=62692) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1080.284864] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-efac3b0a-0872-4fcd-b4da-80753318f032 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249042', 'volume_id': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'name': 'volume-ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e8f645da-3b4a-4f21-9a17-aebdc8d112df', 'attached_at': '', 'detached_at': '', 'volume_id': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'serial': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1080.285756] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-695eba68-9be4-4f25-8855-72e49a40876c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.310710] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8a1eaaa-6ac3-4da5-80fc-11cbb2c80a18 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.336317] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-efac3b0a-0872-4fcd-b4da-80753318f032 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] volume-ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60/volume-ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1080.337392] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fce8cc86-9690-4056-aa69-ca77faa218f4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.355439] env[62692]: DEBUG oslo_vmware.api [None req-efac3b0a-0872-4fcd-b4da-80753318f032 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1080.355439] env[62692]: value = "task-1141766" [ 1080.355439] env[62692]: _type = "Task" [ 1080.355439] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.363677] env[62692]: DEBUG oslo_vmware.api [None req-efac3b0a-0872-4fcd-b4da-80753318f032 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141766, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.560201] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.792s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.562888] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.863s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.564695] env[62692]: INFO nova.compute.claims [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1080.597181] env[62692]: INFO nova.scheduler.client.report [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Deleted allocations for instance ca56c2ec-95fc-44fc-856f-e395bda1703c [ 1080.612716] env[62692]: DEBUG oslo_vmware.api [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52e52526-3e0f-f9c5-45c1-839afbcc7605, 'name': SearchDatastore_Task, 'duration_secs': 0.02235} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.613806] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f8be199-a507-4fb9-9bd0-0cd595aca2be {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.619684] env[62692]: DEBUG oslo_vmware.api [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1080.619684] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52c90dba-ab56-39c8-d5da-6d1d616e848f" [ 1080.619684] env[62692]: _type = "Task" [ 1080.619684] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.632328] env[62692]: DEBUG oslo_vmware.api [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52c90dba-ab56-39c8-d5da-6d1d616e848f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.697029] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Releasing lock "refresh_cache-cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1080.697154] env[62692]: DEBUG nova.compute.manager [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Instance network_info: |[{"id": "a3885c51-6f7b-41e8-8948-41fac442a288", "address": "fa:16:3e:e4:9b:6f", "network": {"id": "f70727d6-a9aa-44b0-a244-2533b3f96571", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-284105837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1037e5d436bd429391159b87a712f6e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3885c51-6f", "ovs_interfaceid": "a3885c51-6f7b-41e8-8948-41fac442a288", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1080.697525] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e4:9b:6f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f44b2fa3-6730-4b87-8839-947eff21213f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a3885c51-6f7b-41e8-8948-41fac442a288', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1080.705575] env[62692]: DEBUG oslo.service.loopingcall [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1080.705829] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1080.706202] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-23f5bdab-00e1-4ed3-ae0c-ce9c82cc86cb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.730033] env[62692]: DEBUG oslo_concurrency.lockutils [None req-84b3c9d5-bce5-416a-9561-94ae737dd0a5 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.730334] env[62692]: DEBUG oslo_concurrency.lockutils [None req-84b3c9d5-bce5-416a-9561-94ae737dd0a5 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.730618] env[62692]: DEBUG nova.compute.manager [None req-84b3c9d5-bce5-416a-9561-94ae737dd0a5 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Going to confirm migration 4 {{(pid=62692) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1080.736090] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1080.736090] env[62692]: value = "task-1141767" [ 1080.736090] env[62692]: _type = "Task" [ 1080.736090] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.736478] env[62692]: DEBUG oslo_vmware.api [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141765, 'name': PowerOnVM_Task, 'duration_secs': 1.268404} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.736916] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1080.737155] env[62692]: DEBUG nova.compute.manager [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1080.740676] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb384ee6-fabe-47a9-ab80-3d01a5ebbf75 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.749462] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141767, 'name': CreateVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.794820] env[62692]: DEBUG nova.compute.manager [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1080.813528] env[62692]: DEBUG nova.compute.manager [req-3ebead8f-1487-44c1-a55f-b333dbb1c695 req-d45561b2-9cb4-422d-9552-b67532046a97 service nova] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Received event network-changed-a3885c51-6f7b-41e8-8948-41fac442a288 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1080.813740] env[62692]: DEBUG nova.compute.manager [req-3ebead8f-1487-44c1-a55f-b333dbb1c695 req-d45561b2-9cb4-422d-9552-b67532046a97 service nova] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Refreshing instance network info cache due to event network-changed-a3885c51-6f7b-41e8-8948-41fac442a288. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1080.813954] env[62692]: DEBUG oslo_concurrency.lockutils [req-3ebead8f-1487-44c1-a55f-b333dbb1c695 req-d45561b2-9cb4-422d-9552-b67532046a97 service nova] Acquiring lock "refresh_cache-cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1080.815305] env[62692]: DEBUG oslo_concurrency.lockutils [req-3ebead8f-1487-44c1-a55f-b333dbb1c695 req-d45561b2-9cb4-422d-9552-b67532046a97 service nova] Acquired lock "refresh_cache-cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.815540] env[62692]: DEBUG nova.network.neutron [req-3ebead8f-1487-44c1-a55f-b333dbb1c695 req-d45561b2-9cb4-422d-9552-b67532046a97 service nova] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Refreshing network info cache for port a3885c51-6f7b-41e8-8948-41fac442a288 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1080.826494] env[62692]: DEBUG nova.virt.hardware [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1080.826811] env[62692]: DEBUG nova.virt.hardware [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1080.827015] env[62692]: DEBUG nova.virt.hardware [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1080.827274] env[62692]: DEBUG nova.virt.hardware [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1080.827427] env[62692]: DEBUG nova.virt.hardware [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1080.827654] env[62692]: DEBUG nova.virt.hardware [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1080.827895] env[62692]: DEBUG nova.virt.hardware [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1080.828081] env[62692]: DEBUG nova.virt.hardware [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1080.828266] env[62692]: DEBUG nova.virt.hardware [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1080.828454] env[62692]: DEBUG nova.virt.hardware [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1080.828645] env[62692]: DEBUG nova.virt.hardware [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1080.829974] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74b59c4e-6c6e-45eb-9aef-e83987856a4a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.839298] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fb40974-4d59-4469-bfc6-b85404367c01 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.863829] env[62692]: DEBUG oslo_vmware.api [None req-efac3b0a-0872-4fcd-b4da-80753318f032 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141766, 'name': ReconfigVM_Task, 'duration_secs': 0.392883} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.864144] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-efac3b0a-0872-4fcd-b4da-80753318f032 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Reconfigured VM instance instance-00000059 to attach disk [datastore2] volume-ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60/volume-ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1080.868885] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-32cbefc8-f4ba-467d-8883-3791ddf8c52a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.884879] env[62692]: DEBUG oslo_vmware.api [None req-efac3b0a-0872-4fcd-b4da-80753318f032 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1080.884879] env[62692]: value = "task-1141768" [ 1080.884879] env[62692]: _type = "Task" [ 1080.884879] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.897456] env[62692]: DEBUG oslo_vmware.api [None req-efac3b0a-0872-4fcd-b4da-80753318f032 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141768, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.967846] env[62692]: DEBUG nova.network.neutron [req-a58d989f-ea55-4afc-84f3-755fcbc31b2a req-a8f3581d-445b-4723-8631-b8478b50ab99 service nova] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Updated VIF entry in instance network info cache for port c3aed689-318a-45c3-a236-6ca6b1bace9b. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1080.968384] env[62692]: DEBUG nova.network.neutron [req-a58d989f-ea55-4afc-84f3-755fcbc31b2a req-a8f3581d-445b-4723-8631-b8478b50ab99 service nova] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Updating instance_info_cache with network_info: [{"id": "c3aed689-318a-45c3-a236-6ca6b1bace9b", "address": "fa:16:3e:23:aa:61", "network": {"id": "e9c538a3-7723-42cb-85b7-5a1597e49cef", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1878627765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71b00c566cea4b21811ea1e1ef5ba0f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc3aed689-31", "ovs_interfaceid": "c3aed689-318a-45c3-a236-6ca6b1bace9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.105952] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b008fdd3-88b0-40bc-bc2a-dce2c7ea1914 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "ca56c2ec-95fc-44fc-856f-e395bda1703c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.477s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.130300] env[62692]: DEBUG oslo_vmware.api [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52c90dba-ab56-39c8-d5da-6d1d616e848f, 'name': SearchDatastore_Task, 'duration_secs': 0.011179} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.130593] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.130867] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 94a07505-661f-4595-9024-8a9da9f2c2fc/94a07505-661f-4595-9024-8a9da9f2c2fc.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1081.131170] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-381c9068-49f2-4343-bddc-cd9714b7f684 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.137696] env[62692]: DEBUG oslo_vmware.api [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1081.137696] env[62692]: value = "task-1141769" [ 1081.137696] env[62692]: _type = "Task" [ 1081.137696] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.146284] env[62692]: DEBUG oslo_vmware.api [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141769, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.244361] env[62692]: DEBUG nova.network.neutron [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Successfully updated port: 06a8d751-3ffa-4fb1-8017-a12d81fc1d98 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1081.259269] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141767, 'name': CreateVM_Task, 'duration_secs': 0.339785} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.259269] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1081.260036] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.260036] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.260970] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1081.261124] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-129a9049-de24-45ad-b5fb-01805f1a603f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.267815] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.271623] env[62692]: DEBUG oslo_vmware.api [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1081.271623] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]523459e6-41d4-d3bb-944b-db942b93224b" [ 1081.271623] env[62692]: _type = "Task" [ 1081.271623] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.281140] env[62692]: DEBUG oslo_vmware.api [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]523459e6-41d4-d3bb-944b-db942b93224b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.284870] env[62692]: DEBUG oslo_concurrency.lockutils [None req-84b3c9d5-bce5-416a-9561-94ae737dd0a5 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "refresh_cache-9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.285058] env[62692]: DEBUG oslo_concurrency.lockutils [None req-84b3c9d5-bce5-416a-9561-94ae737dd0a5 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired lock "refresh_cache-9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.285240] env[62692]: DEBUG nova.network.neutron [None req-84b3c9d5-bce5-416a-9561-94ae737dd0a5 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1081.285426] env[62692]: DEBUG nova.objects.instance [None req-84b3c9d5-bce5-416a-9561-94ae737dd0a5 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lazy-loading 'info_cache' on Instance uuid 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1081.400293] env[62692]: DEBUG oslo_vmware.api [None req-efac3b0a-0872-4fcd-b4da-80753318f032 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141768, 'name': ReconfigVM_Task, 'duration_secs': 0.178793} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.400293] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-efac3b0a-0872-4fcd-b4da-80753318f032 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249042', 'volume_id': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'name': 'volume-ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e8f645da-3b4a-4f21-9a17-aebdc8d112df', 'attached_at': '', 'detached_at': '', 'volume_id': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'serial': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1081.470893] env[62692]: DEBUG oslo_concurrency.lockutils [req-a58d989f-ea55-4afc-84f3-755fcbc31b2a req-a8f3581d-445b-4723-8631-b8478b50ab99 service nova] Releasing lock "refresh_cache-a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.568099] env[62692]: DEBUG nova.network.neutron [req-3ebead8f-1487-44c1-a55f-b333dbb1c695 req-d45561b2-9cb4-422d-9552-b67532046a97 service nova] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Updated VIF entry in instance network info cache for port a3885c51-6f7b-41e8-8948-41fac442a288. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1081.568500] env[62692]: DEBUG nova.network.neutron [req-3ebead8f-1487-44c1-a55f-b333dbb1c695 req-d45561b2-9cb4-422d-9552-b67532046a97 service nova] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Updating instance_info_cache with network_info: [{"id": "a3885c51-6f7b-41e8-8948-41fac442a288", "address": "fa:16:3e:e4:9b:6f", "network": {"id": "f70727d6-a9aa-44b0-a244-2533b3f96571", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-284105837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1037e5d436bd429391159b87a712f6e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3885c51-6f", "ovs_interfaceid": "a3885c51-6f7b-41e8-8948-41fac442a288", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.647600] env[62692]: DEBUG oslo_vmware.api [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141769, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.730620] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7de98d4-12cd-49e4-b088-c02f92c6e0cf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.738415] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40af893a-663d-427c-9aa9-7d26aed1e01e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.771557] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Acquiring lock "refresh_cache-f27e3a4f-d109-472b-a348-5099bdc0e7c8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.771779] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Acquired lock "refresh_cache-f27e3a4f-d109-472b-a348-5099bdc0e7c8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.771940] env[62692]: DEBUG nova.network.neutron [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1081.774098] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0069058-38c2-4d2a-a495-529173371989 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.790330] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5943354b-1b9e-4899-9b1c-24d46a364817 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.794723] env[62692]: DEBUG oslo_vmware.api [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]523459e6-41d4-d3bb-944b-db942b93224b, 'name': SearchDatastore_Task, 'duration_secs': 0.011589} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.797740] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.798008] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1081.798262] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.798457] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.798645] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1081.799589] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-33ca138b-10b5-4033-a954-00054564e4aa {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.812539] env[62692]: DEBUG nova.compute.provider_tree [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1081.815263] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1081.815263] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1081.816333] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b28bf3c-8c5a-4c52-b690-e1d8f61c62f6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.822559] env[62692]: DEBUG oslo_vmware.api [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1081.822559] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52482c11-3e06-3b9d-b461-a11fe0961238" [ 1081.822559] env[62692]: _type = "Task" [ 1081.822559] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.832299] env[62692]: DEBUG oslo_vmware.api [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52482c11-3e06-3b9d-b461-a11fe0961238, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.850211] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Acquiring lock "e086d79d-29e6-41bf-a139-0b680ca0021d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.850501] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Lock "e086d79d-29e6-41bf-a139-0b680ca0021d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.850775] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Acquiring lock "e086d79d-29e6-41bf-a139-0b680ca0021d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.850971] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Lock "e086d79d-29e6-41bf-a139-0b680ca0021d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.851165] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Lock "e086d79d-29e6-41bf-a139-0b680ca0021d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.853548] env[62692]: INFO nova.compute.manager [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Terminating instance [ 1081.855577] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Acquiring lock "refresh_cache-e086d79d-29e6-41bf-a139-0b680ca0021d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.855795] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Acquired lock "refresh_cache-e086d79d-29e6-41bf-a139-0b680ca0021d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.855982] env[62692]: DEBUG nova.network.neutron [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1082.071451] env[62692]: DEBUG oslo_concurrency.lockutils [req-3ebead8f-1487-44c1-a55f-b333dbb1c695 req-d45561b2-9cb4-422d-9552-b67532046a97 service nova] Releasing lock "refresh_cache-cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1082.148123] env[62692]: DEBUG oslo_vmware.api [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141769, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.308021] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "1d9271aa-d41d-48cd-a879-3306e7212197" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.308319] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "1d9271aa-d41d-48cd-a879-3306e7212197" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.311862] env[62692]: DEBUG nova.network.neutron [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1082.317838] env[62692]: DEBUG nova.scheduler.client.report [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1082.339303] env[62692]: DEBUG oslo_vmware.api [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52482c11-3e06-3b9d-b461-a11fe0961238, 'name': SearchDatastore_Task, 'duration_secs': 0.061138} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.340516] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23a85685-e9ff-4d80-8766-1420888b2349 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.346350] env[62692]: DEBUG oslo_vmware.api [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1082.346350] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52ffafe0-58a6-142e-1d47-efc0e63cdbcf" [ 1082.346350] env[62692]: _type = "Task" [ 1082.346350] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.355406] env[62692]: DEBUG oslo_vmware.api [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52ffafe0-58a6-142e-1d47-efc0e63cdbcf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.386855] env[62692]: DEBUG nova.network.neutron [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1082.446962] env[62692]: DEBUG nova.objects.instance [None req-efac3b0a-0872-4fcd-b4da-80753318f032 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lazy-loading 'flavor' on Instance uuid e8f645da-3b4a-4f21-9a17-aebdc8d112df {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1082.473166] env[62692]: DEBUG nova.network.neutron [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.509191] env[62692]: DEBUG nova.network.neutron [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Updating instance_info_cache with network_info: [{"id": "06a8d751-3ffa-4fb1-8017-a12d81fc1d98", "address": "fa:16:3e:cc:bd:2b", "network": {"id": "e25b4084-b168-4a95-af81-4141dfb9e1c1", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1332280091-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43796ce588804516bdbbba3ed5b0e25f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f925dc8-2145-457e-a4d4-c07117356dd0", "external-id": "nsx-vlan-transportzone-356", "segmentation_id": 356, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06a8d751-3f", "ovs_interfaceid": "06a8d751-3ffa-4fb1-8017-a12d81fc1d98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.649428] env[62692]: DEBUG oslo_vmware.api [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141769, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.026009} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.649719] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 94a07505-661f-4595-9024-8a9da9f2c2fc/94a07505-661f-4595-9024-8a9da9f2c2fc.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1082.649993] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1082.650202] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-105b64e4-b1eb-403c-a810-cb4dc0a3423d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.656515] env[62692]: DEBUG oslo_vmware.api [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1082.656515] env[62692]: value = "task-1141770" [ 1082.656515] env[62692]: _type = "Task" [ 1082.656515] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.664294] env[62692]: DEBUG oslo_vmware.api [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141770, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.738423] env[62692]: DEBUG nova.network.neutron [None req-84b3c9d5-bce5-416a-9561-94ae737dd0a5 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Updating instance_info_cache with network_info: [{"id": "d5ff0422-e9b3-455c-bd28-422096794288", "address": "fa:16:3e:46:38:ff", "network": {"id": "f6352999-f943-431d-b54a-c9dc1d810a85", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-121037111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9cb8ae08824245a7911de7455e7fcb68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c842425c-544e-4ce2-9657-512723bd318e", "external-id": "nsx-vlan-transportzone-80", "segmentation_id": 80, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5ff0422-e9", "ovs_interfaceid": "d5ff0422-e9b3-455c-bd28-422096794288", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.810742] env[62692]: DEBUG nova.compute.manager [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1082.823807] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.261s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.824404] env[62692]: DEBUG nova.compute.manager [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1082.827353] env[62692]: DEBUG oslo_concurrency.lockutils [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.026s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.827635] env[62692]: DEBUG nova.objects.instance [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Lazy-loading 'resources' on Instance uuid 1b753882-9118-479c-aa3e-cbb8bdf2e086 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1082.838010] env[62692]: DEBUG nova.compute.manager [req-130cdce6-b691-4070-961f-464d98bdc15c req-ebc60de5-e9ef-427e-95ed-800833d6f20f service nova] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Received event network-vif-plugged-06a8d751-3ffa-4fb1-8017-a12d81fc1d98 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1082.838418] env[62692]: DEBUG oslo_concurrency.lockutils [req-130cdce6-b691-4070-961f-464d98bdc15c req-ebc60de5-e9ef-427e-95ed-800833d6f20f service nova] Acquiring lock "f27e3a4f-d109-472b-a348-5099bdc0e7c8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.838418] env[62692]: DEBUG oslo_concurrency.lockutils [req-130cdce6-b691-4070-961f-464d98bdc15c req-ebc60de5-e9ef-427e-95ed-800833d6f20f service nova] Lock "f27e3a4f-d109-472b-a348-5099bdc0e7c8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.839766] env[62692]: DEBUG oslo_concurrency.lockutils [req-130cdce6-b691-4070-961f-464d98bdc15c req-ebc60de5-e9ef-427e-95ed-800833d6f20f service nova] Lock "f27e3a4f-d109-472b-a348-5099bdc0e7c8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.839766] env[62692]: DEBUG nova.compute.manager [req-130cdce6-b691-4070-961f-464d98bdc15c req-ebc60de5-e9ef-427e-95ed-800833d6f20f service nova] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] No waiting events found dispatching network-vif-plugged-06a8d751-3ffa-4fb1-8017-a12d81fc1d98 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1082.839766] env[62692]: WARNING nova.compute.manager [req-130cdce6-b691-4070-961f-464d98bdc15c req-ebc60de5-e9ef-427e-95ed-800833d6f20f service nova] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Received unexpected event network-vif-plugged-06a8d751-3ffa-4fb1-8017-a12d81fc1d98 for instance with vm_state building and task_state spawning. [ 1082.839766] env[62692]: DEBUG nova.compute.manager [req-130cdce6-b691-4070-961f-464d98bdc15c req-ebc60de5-e9ef-427e-95ed-800833d6f20f service nova] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Received event network-changed-06a8d751-3ffa-4fb1-8017-a12d81fc1d98 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1082.839766] env[62692]: DEBUG nova.compute.manager [req-130cdce6-b691-4070-961f-464d98bdc15c req-ebc60de5-e9ef-427e-95ed-800833d6f20f service nova] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Refreshing instance network info cache due to event network-changed-06a8d751-3ffa-4fb1-8017-a12d81fc1d98. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1082.839766] env[62692]: DEBUG oslo_concurrency.lockutils [req-130cdce6-b691-4070-961f-464d98bdc15c req-ebc60de5-e9ef-427e-95ed-800833d6f20f service nova] Acquiring lock "refresh_cache-f27e3a4f-d109-472b-a348-5099bdc0e7c8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.857499] env[62692]: DEBUG oslo_vmware.api [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52ffafe0-58a6-142e-1d47-efc0e63cdbcf, 'name': SearchDatastore_Task, 'duration_secs': 0.023497} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.857824] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1082.858167] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43/cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1082.858903] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-559db0a4-b1cc-42ab-98a7-020eee3cc952 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.866548] env[62692]: DEBUG oslo_vmware.api [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1082.866548] env[62692]: value = "task-1141771" [ 1082.866548] env[62692]: _type = "Task" [ 1082.866548] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.874845] env[62692]: DEBUG oslo_vmware.api [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141771, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.956447] env[62692]: DEBUG oslo_concurrency.lockutils [None req-efac3b0a-0872-4fcd-b4da-80753318f032 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "e8f645da-3b4a-4f21-9a17-aebdc8d112df" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.275s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.977066] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Releasing lock "refresh_cache-e086d79d-29e6-41bf-a139-0b680ca0021d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1082.977515] env[62692]: DEBUG nova.compute.manager [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1082.977718] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1082.979106] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70960a36-dd79-4437-ad1f-b602525c27cb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.987572] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1082.988337] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f99b19be-9593-43cd-895c-bccf1181b9fd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.994730] env[62692]: DEBUG oslo_vmware.api [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Waiting for the task: (returnval){ [ 1082.994730] env[62692]: value = "task-1141772" [ 1082.994730] env[62692]: _type = "Task" [ 1082.994730] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.004197] env[62692]: DEBUG oslo_vmware.api [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141772, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.012194] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Releasing lock "refresh_cache-f27e3a4f-d109-472b-a348-5099bdc0e7c8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.013223] env[62692]: DEBUG nova.compute.manager [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Instance network_info: |[{"id": "06a8d751-3ffa-4fb1-8017-a12d81fc1d98", "address": "fa:16:3e:cc:bd:2b", "network": {"id": "e25b4084-b168-4a95-af81-4141dfb9e1c1", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1332280091-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43796ce588804516bdbbba3ed5b0e25f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f925dc8-2145-457e-a4d4-c07117356dd0", "external-id": "nsx-vlan-transportzone-356", "segmentation_id": 356, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06a8d751-3f", "ovs_interfaceid": "06a8d751-3ffa-4fb1-8017-a12d81fc1d98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1083.014093] env[62692]: DEBUG oslo_concurrency.lockutils [req-130cdce6-b691-4070-961f-464d98bdc15c req-ebc60de5-e9ef-427e-95ed-800833d6f20f service nova] Acquired lock "refresh_cache-f27e3a4f-d109-472b-a348-5099bdc0e7c8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.014180] env[62692]: DEBUG nova.network.neutron [req-130cdce6-b691-4070-961f-464d98bdc15c req-ebc60de5-e9ef-427e-95ed-800833d6f20f service nova] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Refreshing network info cache for port 06a8d751-3ffa-4fb1-8017-a12d81fc1d98 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1083.015782] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cc:bd:2b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0f925dc8-2145-457e-a4d4-c07117356dd0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '06a8d751-3ffa-4fb1-8017-a12d81fc1d98', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1083.023898] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Creating folder: Project (43796ce588804516bdbbba3ed5b0e25f). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1083.031431] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1aaacdc3-8ffc-42a7-bb80-c42bfa1f7866 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.042293] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Created folder: Project (43796ce588804516bdbbba3ed5b0e25f) in parent group-v248868. [ 1083.042421] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Creating folder: Instances. Parent ref: group-v249045. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1083.043059] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b800fdcd-fb65-4371-aa89-70478959e3a4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.051617] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Created folder: Instances in parent group-v249045. [ 1083.052206] env[62692]: DEBUG oslo.service.loopingcall [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1083.052206] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1083.052367] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-852d57b6-c3dd-4427-b814-9b115a005b4e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.072519] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1083.072519] env[62692]: value = "task-1141775" [ 1083.072519] env[62692]: _type = "Task" [ 1083.072519] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.082648] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141775, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.167280] env[62692]: DEBUG oslo_vmware.api [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141770, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.214913} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.167630] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1083.168398] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7b12a5b-3f87-41bc-ac66-0f746e9d5843 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.191170] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] 94a07505-661f-4595-9024-8a9da9f2c2fc/94a07505-661f-4595-9024-8a9da9f2c2fc.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1083.194112] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7c4039de-3c4c-4fe2-90b5-7fc9253640f8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.215598] env[62692]: DEBUG oslo_vmware.api [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1083.215598] env[62692]: value = "task-1141776" [ 1083.215598] env[62692]: _type = "Task" [ 1083.215598] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.225913] env[62692]: DEBUG oslo_vmware.api [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141776, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.241944] env[62692]: DEBUG oslo_concurrency.lockutils [None req-84b3c9d5-bce5-416a-9561-94ae737dd0a5 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Releasing lock "refresh_cache-9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.242296] env[62692]: DEBUG nova.objects.instance [None req-84b3c9d5-bce5-416a-9561-94ae737dd0a5 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lazy-loading 'migration_context' on Instance uuid 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1083.284609] env[62692]: DEBUG nova.network.neutron [req-130cdce6-b691-4070-961f-464d98bdc15c req-ebc60de5-e9ef-427e-95ed-800833d6f20f service nova] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Updated VIF entry in instance network info cache for port 06a8d751-3ffa-4fb1-8017-a12d81fc1d98. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1083.284997] env[62692]: DEBUG nova.network.neutron [req-130cdce6-b691-4070-961f-464d98bdc15c req-ebc60de5-e9ef-427e-95ed-800833d6f20f service nova] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Updating instance_info_cache with network_info: [{"id": "06a8d751-3ffa-4fb1-8017-a12d81fc1d98", "address": "fa:16:3e:cc:bd:2b", "network": {"id": "e25b4084-b168-4a95-af81-4141dfb9e1c1", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1332280091-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43796ce588804516bdbbba3ed5b0e25f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f925dc8-2145-457e-a4d4-c07117356dd0", "external-id": "nsx-vlan-transportzone-356", "segmentation_id": 356, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06a8d751-3f", "ovs_interfaceid": "06a8d751-3ffa-4fb1-8017-a12d81fc1d98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1083.331212] env[62692]: DEBUG nova.compute.utils [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1083.336694] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.337060] env[62692]: DEBUG nova.compute.manager [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1083.337528] env[62692]: DEBUG nova.network.neutron [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1083.378797] env[62692]: DEBUG oslo_vmware.api [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141771, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.414549] env[62692]: DEBUG nova.policy [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f2e2ee26163841ddb15bb67520e98394', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e13f8888d5e84eed8247101558e29a56', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1083.508518] env[62692]: DEBUG oslo_vmware.api [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141772, 'name': PowerOffVM_Task, 'duration_secs': 0.125002} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.508518] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1083.508518] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1083.508518] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-867f0045-d2b2-44a7-97a5-27354d009764 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.532932] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1083.533219] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1083.533403] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Deleting the datastore file [datastore2] e086d79d-29e6-41bf-a139-0b680ca0021d {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1083.534084] env[62692]: INFO nova.compute.manager [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Rebuilding instance [ 1083.535881] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9e2c3cc2-ca9a-4a67-9b60-5d4343f9c719 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.541723] env[62692]: DEBUG oslo_vmware.api [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Waiting for the task: (returnval){ [ 1083.541723] env[62692]: value = "task-1141778" [ 1083.541723] env[62692]: _type = "Task" [ 1083.541723] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.552383] env[62692]: DEBUG oslo_vmware.api [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141778, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.562769] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fa095b3-2719-4ab2-8af1-0da3dc27429e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.569301] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-220c0324-ff42-4450-876b-92f6921b81cc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.603686] env[62692]: DEBUG nova.compute.manager [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1083.604656] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fc117a4-4f86-4c84-aa4a-e000f1b824dc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.607867] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5ca0628-5b8e-4802-b30f-f5bf4fcdd945 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.613771] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141775, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.624066] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ba63eae-c6c7-4528-932b-4bf0a94682d8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.636109] env[62692]: DEBUG nova.compute.provider_tree [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1083.726792] env[62692]: DEBUG oslo_vmware.api [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141776, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.748100] env[62692]: DEBUG nova.objects.base [None req-84b3c9d5-bce5-416a-9561-94ae737dd0a5 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Object Instance<9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da> lazy-loaded attributes: info_cache,migration_context {{(pid=62692) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1083.748100] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fda0652-5531-413e-8491-3cb06ee07a80 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.751175] env[62692]: DEBUG nova.network.neutron [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Successfully created port: 3c66a7e2-2738-4873-b748-348237876eda {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1083.771963] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9469c622-4fd0-4d2d-808b-9cb2756cdab5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.777380] env[62692]: DEBUG oslo_vmware.api [None req-84b3c9d5-bce5-416a-9561-94ae737dd0a5 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1083.777380] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52fc53ec-9144-84b8-5245-a0f54c8d9e0d" [ 1083.777380] env[62692]: _type = "Task" [ 1083.777380] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.784866] env[62692]: DEBUG oslo_vmware.api [None req-84b3c9d5-bce5-416a-9561-94ae737dd0a5 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52fc53ec-9144-84b8-5245-a0f54c8d9e0d, 'name': SearchDatastore_Task} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.785156] env[62692]: DEBUG oslo_concurrency.lockutils [None req-84b3c9d5-bce5-416a-9561-94ae737dd0a5 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.787752] env[62692]: DEBUG oslo_concurrency.lockutils [req-130cdce6-b691-4070-961f-464d98bdc15c req-ebc60de5-e9ef-427e-95ed-800833d6f20f service nova] Releasing lock "refresh_cache-f27e3a4f-d109-472b-a348-5099bdc0e7c8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.836856] env[62692]: DEBUG nova.compute.manager [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1083.879531] env[62692]: DEBUG oslo_vmware.api [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141771, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.638852} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.879893] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43/cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1083.880071] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1083.880235] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-88fdcb9e-14b6-4f32-a5cb-955c1fad947d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.887586] env[62692]: DEBUG oslo_vmware.api [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1083.887586] env[62692]: value = "task-1141779" [ 1083.887586] env[62692]: _type = "Task" [ 1083.887586] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.896069] env[62692]: DEBUG oslo_vmware.api [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141779, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.051602] env[62692]: DEBUG oslo_vmware.api [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Task: {'id': task-1141778, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.104172} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.051892] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1084.052107] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1084.052296] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1084.052472] env[62692]: INFO nova.compute.manager [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Took 1.07 seconds to destroy the instance on the hypervisor. [ 1084.052718] env[62692]: DEBUG oslo.service.loopingcall [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1084.052932] env[62692]: DEBUG nova.compute.manager [-] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1084.053052] env[62692]: DEBUG nova.network.neutron [-] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1084.069182] env[62692]: DEBUG nova.network.neutron [-] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1084.085802] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141775, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.122929] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1084.122929] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-14dfcfae-17b6-4131-b0ed-5b7ac3b032eb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.130997] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1084.130997] env[62692]: value = "task-1141780" [ 1084.130997] env[62692]: _type = "Task" [ 1084.130997] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.139248] env[62692]: DEBUG nova.scheduler.client.report [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1084.142237] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141780, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.226381] env[62692]: DEBUG oslo_vmware.api [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141776, 'name': ReconfigVM_Task, 'duration_secs': 0.516028} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.226663] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Reconfigured VM instance instance-00000061 to attach disk [datastore2] 94a07505-661f-4595-9024-8a9da9f2c2fc/94a07505-661f-4595-9024-8a9da9f2c2fc.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1084.227279] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-69afb344-8b00-438d-bc50-3a4460f61910 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.232905] env[62692]: DEBUG oslo_vmware.api [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1084.232905] env[62692]: value = "task-1141781" [ 1084.232905] env[62692]: _type = "Task" [ 1084.232905] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.240398] env[62692]: DEBUG oslo_vmware.api [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141781, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.398132] env[62692]: DEBUG oslo_vmware.api [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141779, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066258} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.398492] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1084.399182] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f798113b-bd4d-4869-b063-6a9f5040bdd6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.421090] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43/cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1084.421656] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0305a165-785e-40d0-896f-c7251c0e2a8a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.441328] env[62692]: DEBUG oslo_vmware.api [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1084.441328] env[62692]: value = "task-1141782" [ 1084.441328] env[62692]: _type = "Task" [ 1084.441328] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.450845] env[62692]: DEBUG oslo_vmware.api [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141782, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.571612] env[62692]: DEBUG nova.network.neutron [-] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.587023] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141775, 'name': CreateVM_Task, 'duration_secs': 1.384903} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.587416] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1084.587943] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.588149] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.588527] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1084.588740] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ccd5779-0dac-42df-969d-4e8e8b36e406 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.593241] env[62692]: DEBUG oslo_vmware.api [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Waiting for the task: (returnval){ [ 1084.593241] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5227abbc-f66d-e5e2-b8e0-aa990a55891c" [ 1084.593241] env[62692]: _type = "Task" [ 1084.593241] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.602030] env[62692]: DEBUG oslo_vmware.api [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5227abbc-f66d-e5e2-b8e0-aa990a55891c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.640923] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141780, 'name': PowerOffVM_Task, 'duration_secs': 0.452862} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.640923] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1084.644042] env[62692]: DEBUG oslo_concurrency.lockutils [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.817s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.647999] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 7.146s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.648179] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.648306] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62692) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1084.648611] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 3.381s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.648891] env[62692]: DEBUG nova.objects.instance [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62692) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1084.653899] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23762220-c1e7-4f7e-b3aa-3f7d85d17052 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.664582] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e44bfd92-8940-4795-8655-744fcb58e5de {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.670818] env[62692]: INFO nova.scheduler.client.report [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Deleted allocations for instance 1b753882-9118-479c-aa3e-cbb8bdf2e086 [ 1084.685976] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea3fc410-43c1-4da4-a8fd-1985c87e082f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.694469] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d09f28f-23e7-4777-9e6c-24e9aae757c5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.727624] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180618MB free_disk=166GB free_vcpus=48 pci_devices=None {{(pid=62692) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1084.727789] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.728989] env[62692]: INFO nova.compute.manager [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Detaching volume ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60 [ 1084.745201] env[62692]: DEBUG oslo_vmware.api [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141781, 'name': Rename_Task, 'duration_secs': 0.154659} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.745539] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1084.746811] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-288933a0-1812-4dc6-b3c7-438463b6b80f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.754916] env[62692]: DEBUG oslo_vmware.api [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1084.754916] env[62692]: value = "task-1141783" [ 1084.754916] env[62692]: _type = "Task" [ 1084.754916] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.766873] env[62692]: INFO nova.virt.block_device [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Attempting to driver detach volume ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60 from mountpoint /dev/sdb [ 1084.767180] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Volume detach. Driver type: vmdk {{(pid=62692) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1084.767387] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249042', 'volume_id': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'name': 'volume-ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e8f645da-3b4a-4f21-9a17-aebdc8d112df', 'attached_at': '', 'detached_at': '', 'volume_id': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'serial': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1084.768259] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd68c0b-8af1-42e9-ae12-db2753cc62cd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.776051] env[62692]: DEBUG oslo_vmware.api [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141783, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.794579] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53a49329-1fee-45dd-825b-c955af28bdb2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.802217] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8c260e4-d934-44f1-a314-4ba0e1b434f6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.823138] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94b4e791-2891-4257-a190-f30a0d226b55 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.839669] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] The volume has not been displaced from its original location: [datastore2] volume-ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60/volume-ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60.vmdk. No consolidation needed. {{(pid=62692) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1084.844984] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Reconfiguring VM instance instance-00000059 to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1084.845345] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8498e7aa-8edb-42f2-9263-87099e092690 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.858833] env[62692]: DEBUG nova.compute.manager [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1084.866795] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1084.866795] env[62692]: value = "task-1141784" [ 1084.866795] env[62692]: _type = "Task" [ 1084.866795] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.875643] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141784, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.887694] env[62692]: DEBUG nova.virt.hardware [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1084.888031] env[62692]: DEBUG nova.virt.hardware [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1084.888227] env[62692]: DEBUG nova.virt.hardware [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1084.888447] env[62692]: DEBUG nova.virt.hardware [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1084.888642] env[62692]: DEBUG nova.virt.hardware [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1084.888805] env[62692]: DEBUG nova.virt.hardware [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1084.889055] env[62692]: DEBUG nova.virt.hardware [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1084.889286] env[62692]: DEBUG nova.virt.hardware [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1084.889546] env[62692]: DEBUG nova.virt.hardware [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1084.889759] env[62692]: DEBUG nova.virt.hardware [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1084.890118] env[62692]: DEBUG nova.virt.hardware [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1084.891017] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b84e1745-00cb-494b-a9f1-9bba101e69d9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.898885] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9bbeb4b-4e83-4f96-b92f-fc1b3bf6e09b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.953112] env[62692]: DEBUG oslo_vmware.api [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141782, 'name': ReconfigVM_Task, 'duration_secs': 0.335707} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.953403] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Reconfigured VM instance instance-00000062 to attach disk [datastore2] cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43/cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1084.954094] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-226c2530-bf7a-4cb2-a4eb-67113b6a1ecb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.960562] env[62692]: DEBUG oslo_vmware.api [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1084.960562] env[62692]: value = "task-1141785" [ 1084.960562] env[62692]: _type = "Task" [ 1084.960562] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.969198] env[62692]: DEBUG oslo_vmware.api [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141785, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.074661] env[62692]: INFO nova.compute.manager [-] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Took 1.02 seconds to deallocate network for instance. [ 1085.107870] env[62692]: DEBUG oslo_vmware.api [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5227abbc-f66d-e5e2-b8e0-aa990a55891c, 'name': SearchDatastore_Task, 'duration_secs': 0.010252} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.108281] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.108715] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1085.108990] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.109175] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.109362] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1085.109644] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d1cab95f-b1fd-4aac-95ed-4796626b662b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.117689] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1085.117873] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1085.118632] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e57caa02-bc98-4400-8bf1-b7eb003d3f4e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.124109] env[62692]: DEBUG oslo_vmware.api [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Waiting for the task: (returnval){ [ 1085.124109] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]524533bb-6dd9-4fbd-3717-86784162bd96" [ 1085.124109] env[62692]: _type = "Task" [ 1085.124109] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.131781] env[62692]: DEBUG oslo_vmware.api [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]524533bb-6dd9-4fbd-3717-86784162bd96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.163665] env[62692]: DEBUG nova.compute.manager [req-18f740c0-95ab-4edd-9357-7bb1dbf24625 req-0d439680-94ed-44a8-9df7-3d3e29ab1d56 service nova] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Received event network-vif-plugged-3c66a7e2-2738-4873-b748-348237876eda {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1085.163951] env[62692]: DEBUG oslo_concurrency.lockutils [req-18f740c0-95ab-4edd-9357-7bb1dbf24625 req-0d439680-94ed-44a8-9df7-3d3e29ab1d56 service nova] Acquiring lock "f92fae48-7285-4b69-97c9-2360227c18fc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.164424] env[62692]: DEBUG oslo_concurrency.lockutils [req-18f740c0-95ab-4edd-9357-7bb1dbf24625 req-0d439680-94ed-44a8-9df7-3d3e29ab1d56 service nova] Lock "f92fae48-7285-4b69-97c9-2360227c18fc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.164424] env[62692]: DEBUG oslo_concurrency.lockutils [req-18f740c0-95ab-4edd-9357-7bb1dbf24625 req-0d439680-94ed-44a8-9df7-3d3e29ab1d56 service nova] Lock "f92fae48-7285-4b69-97c9-2360227c18fc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.164528] env[62692]: DEBUG nova.compute.manager [req-18f740c0-95ab-4edd-9357-7bb1dbf24625 req-0d439680-94ed-44a8-9df7-3d3e29ab1d56 service nova] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] No waiting events found dispatching network-vif-plugged-3c66a7e2-2738-4873-b748-348237876eda {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1085.164684] env[62692]: WARNING nova.compute.manager [req-18f740c0-95ab-4edd-9357-7bb1dbf24625 req-0d439680-94ed-44a8-9df7-3d3e29ab1d56 service nova] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Received unexpected event network-vif-plugged-3c66a7e2-2738-4873-b748-348237876eda for instance with vm_state building and task_state spawning. [ 1085.188726] env[62692]: DEBUG oslo_concurrency.lockutils [None req-336067f1-4a8d-4d80-9526-b653b3852fc5 tempest-FloatingIPsAssociationTestJSON-1041883856 tempest-FloatingIPsAssociationTestJSON-1041883856-project-member] Lock "1b753882-9118-479c-aa3e-cbb8bdf2e086" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.320s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.260477] env[62692]: DEBUG nova.network.neutron [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Successfully updated port: 3c66a7e2-2738-4873-b748-348237876eda {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1085.269417] env[62692]: DEBUG oslo_vmware.api [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141783, 'name': PowerOnVM_Task, 'duration_secs': 0.495051} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.269762] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1085.270017] env[62692]: INFO nova.compute.manager [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Took 9.12 seconds to spawn the instance on the hypervisor. [ 1085.270256] env[62692]: DEBUG nova.compute.manager [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1085.271735] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58647827-2c04-4d5f-a8a2-e2c6e4111a10 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.378828] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141784, 'name': ReconfigVM_Task, 'duration_secs': 0.244908} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.379320] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Reconfigured VM instance instance-00000059 to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1085.384822] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-18450e8b-9641-4d16-b147-47b445a0cc9a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.400569] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1085.400569] env[62692]: value = "task-1141786" [ 1085.400569] env[62692]: _type = "Task" [ 1085.400569] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.408940] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141786, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.470108] env[62692]: DEBUG oslo_vmware.api [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141785, 'name': Rename_Task, 'duration_secs': 0.145545} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.470416] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1085.470703] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8bf5548e-d643-4928-acbb-a79bcb057e51 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.476776] env[62692]: DEBUG oslo_vmware.api [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1085.476776] env[62692]: value = "task-1141787" [ 1085.476776] env[62692]: _type = "Task" [ 1085.476776] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.486289] env[62692]: DEBUG oslo_vmware.api [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141787, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.582193] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.634149] env[62692]: DEBUG oslo_vmware.api [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]524533bb-6dd9-4fbd-3717-86784162bd96, 'name': SearchDatastore_Task, 'duration_secs': 0.013353} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.635171] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a49a7270-bab0-4329-96cc-4de4dbe44aad {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.642151] env[62692]: DEBUG oslo_vmware.api [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Waiting for the task: (returnval){ [ 1085.642151] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5289320e-6821-f083-9b67-6016dd7d03c4" [ 1085.642151] env[62692]: _type = "Task" [ 1085.642151] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.651218] env[62692]: DEBUG oslo_vmware.api [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5289320e-6821-f083-9b67-6016dd7d03c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.663964] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8cea33f3-5fe8-4ba3-bb6e-bd473b0a7c0b tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.665138] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.329s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.666714] env[62692]: INFO nova.compute.claims [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1085.763245] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "refresh_cache-f92fae48-7285-4b69-97c9-2360227c18fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.763245] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquired lock "refresh_cache-f92fae48-7285-4b69-97c9-2360227c18fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.763429] env[62692]: DEBUG nova.network.neutron [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1085.791507] env[62692]: INFO nova.compute.manager [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Took 22.89 seconds to build instance. [ 1085.911316] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141786, 'name': ReconfigVM_Task, 'duration_secs': 0.288044} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.911641] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249042', 'volume_id': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'name': 'volume-ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e8f645da-3b4a-4f21-9a17-aebdc8d112df', 'attached_at': '', 'detached_at': '', 'volume_id': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'serial': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1085.991307] env[62692]: DEBUG oslo_vmware.api [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141787, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.152365] env[62692]: DEBUG oslo_vmware.api [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5289320e-6821-f083-9b67-6016dd7d03c4, 'name': SearchDatastore_Task, 'duration_secs': 0.011739} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.152635] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.152902] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] f27e3a4f-d109-472b-a348-5099bdc0e7c8/f27e3a4f-d109-472b-a348-5099bdc0e7c8.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1086.153181] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6a83a858-bdbf-41db-9622-ffc40dc5ae49 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.159649] env[62692]: DEBUG oslo_vmware.api [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Waiting for the task: (returnval){ [ 1086.159649] env[62692]: value = "task-1141788" [ 1086.159649] env[62692]: _type = "Task" [ 1086.159649] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.167216] env[62692]: DEBUG oslo_vmware.api [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Task: {'id': task-1141788, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.293862] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9f6a1fe5-771f-4ee2-97ed-5906ecd8742c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "94a07505-661f-4595-9024-8a9da9f2c2fc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.398s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.301376] env[62692]: DEBUG nova.network.neutron [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1086.448453] env[62692]: DEBUG nova.network.neutron [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Updating instance_info_cache with network_info: [{"id": "3c66a7e2-2738-4873-b748-348237876eda", "address": "fa:16:3e:68:2e:27", "network": {"id": "d0e00428-1d04-4b63-8975-7f3085784358", "bridge": "br-int", "label": "tempest-ServersTestJSON-408035146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e13f8888d5e84eed8247101558e29a56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c66a7e2-27", "ovs_interfaceid": "3c66a7e2-2738-4873-b748-348237876eda", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1086.488676] env[62692]: DEBUG oslo_vmware.api [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141787, 'name': PowerOnVM_Task, 'duration_secs': 0.801613} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.488959] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1086.489184] env[62692]: INFO nova.compute.manager [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Took 8.03 seconds to spawn the instance on the hypervisor. [ 1086.489369] env[62692]: DEBUG nova.compute.manager [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1086.490205] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f9610da-48ad-42f1-914b-856dbbced610 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.658202] env[62692]: DEBUG nova.compute.manager [req-f84cdc2f-cd43-4924-9993-9d1cad623ca0 req-c5b296e2-18bd-4dfd-a0ce-195ec28cfd21 service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Received event network-changed-d51ffd3f-c023-45e2-8bfa-1424b7ccb188 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1086.658316] env[62692]: DEBUG nova.compute.manager [req-f84cdc2f-cd43-4924-9993-9d1cad623ca0 req-c5b296e2-18bd-4dfd-a0ce-195ec28cfd21 service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Refreshing instance network info cache due to event network-changed-d51ffd3f-c023-45e2-8bfa-1424b7ccb188. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1086.658798] env[62692]: DEBUG oslo_concurrency.lockutils [req-f84cdc2f-cd43-4924-9993-9d1cad623ca0 req-c5b296e2-18bd-4dfd-a0ce-195ec28cfd21 service nova] Acquiring lock "refresh_cache-94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.659010] env[62692]: DEBUG oslo_concurrency.lockutils [req-f84cdc2f-cd43-4924-9993-9d1cad623ca0 req-c5b296e2-18bd-4dfd-a0ce-195ec28cfd21 service nova] Acquired lock "refresh_cache-94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.660075] env[62692]: DEBUG nova.network.neutron [req-f84cdc2f-cd43-4924-9993-9d1cad623ca0 req-c5b296e2-18bd-4dfd-a0ce-195ec28cfd21 service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Refreshing network info cache for port d51ffd3f-c023-45e2-8bfa-1424b7ccb188 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1086.683983] env[62692]: DEBUG oslo_vmware.api [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Task: {'id': task-1141788, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.919618] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8824814-8322-4b0a-ab6a-457873357f7b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.929917] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cd19f98-c774-411b-9079-34f83e2dbce1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.971981] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Releasing lock "refresh_cache-f92fae48-7285-4b69-97c9-2360227c18fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.972353] env[62692]: DEBUG nova.compute.manager [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Instance network_info: |[{"id": "3c66a7e2-2738-4873-b748-348237876eda", "address": "fa:16:3e:68:2e:27", "network": {"id": "d0e00428-1d04-4b63-8975-7f3085784358", "bridge": "br-int", "label": "tempest-ServersTestJSON-408035146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e13f8888d5e84eed8247101558e29a56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c66a7e2-27", "ovs_interfaceid": "3c66a7e2-2738-4873-b748-348237876eda", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1086.972918] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1086.973814] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:68:2e:27', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3c66a7e2-2738-4873-b748-348237876eda', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1086.982123] env[62692]: DEBUG oslo.service.loopingcall [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1086.982414] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d37fee22-5bd6-4a90-b4ef-6c61461e43eb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.984809] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2433c27-f853-4d2a-a0b3-49f3b7280001 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.987611] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1086.988213] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-467a998b-316d-4279-8f52-f294f3ecdb6a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.016027] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1087.016027] env[62692]: value = "task-1141789" [ 1087.016027] env[62692]: _type = "Task" [ 1087.016027] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.019261] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e91bd92-b4fb-4892-95b8-b02a2847b841 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.025383] env[62692]: INFO nova.compute.manager [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Took 22.72 seconds to build instance. [ 1087.031272] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1087.031272] env[62692]: value = "task-1141790" [ 1087.031272] env[62692]: _type = "Task" [ 1087.031272] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.045178] env[62692]: DEBUG nova.compute.provider_tree [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1087.047568] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] VM already powered off {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1087.047975] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Volume detach. Driver type: vmdk {{(pid=62692) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1087.048174] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249042', 'volume_id': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'name': 'volume-ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e8f645da-3b4a-4f21-9a17-aebdc8d112df', 'attached_at': '', 'detached_at': '', 'volume_id': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'serial': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1087.049644] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a480c76-be42-4aa8-8273-cf4305d4b6e4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.056992] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141790, 'name': CreateVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.077637] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c3c4b70-cc17-41a6-bf21-46cc3f050749 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.083675] env[62692]: WARNING nova.virt.vmwareapi.driver [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1087.083999] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1087.088020] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7fab1b9-ba6f-4da8-a6f9-307fc44f363d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.092732] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1087.093119] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d67cb6d3-b493-4eb9-ae01-ee6ce10d8d3c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.155526] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1087.155773] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1087.156072] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Deleting the datastore file [datastore1] e8f645da-3b4a-4f21-9a17-aebdc8d112df {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1087.156292] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-72de8803-15ac-4ae2-a1fa-3b1d70946fbb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.166522] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1087.166522] env[62692]: value = "task-1141792" [ 1087.166522] env[62692]: _type = "Task" [ 1087.166522] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.189049] env[62692]: DEBUG oslo_vmware.api [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Task: {'id': task-1141788, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.853387} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.193533] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] f27e3a4f-d109-472b-a348-5099bdc0e7c8/f27e3a4f-d109-472b-a348-5099bdc0e7c8.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1087.193956] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1087.194847] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141792, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.195184] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-320c1534-9296-4ac6-8bab-cbda62b90789 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.202648] env[62692]: DEBUG oslo_vmware.api [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Waiting for the task: (returnval){ [ 1087.202648] env[62692]: value = "task-1141793" [ 1087.202648] env[62692]: _type = "Task" [ 1087.202648] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.214365] env[62692]: DEBUG oslo_vmware.api [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Task: {'id': task-1141793, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.469475] env[62692]: DEBUG nova.compute.manager [req-cc91433d-6ec0-4139-b309-4306411e6906 req-4f2e4a0e-1b87-4d1f-97e9-833d3d87e5a1 service nova] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Received event network-changed-3c66a7e2-2738-4873-b748-348237876eda {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1087.469475] env[62692]: DEBUG nova.compute.manager [req-cc91433d-6ec0-4139-b309-4306411e6906 req-4f2e4a0e-1b87-4d1f-97e9-833d3d87e5a1 service nova] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Refreshing instance network info cache due to event network-changed-3c66a7e2-2738-4873-b748-348237876eda. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1087.469475] env[62692]: DEBUG oslo_concurrency.lockutils [req-cc91433d-6ec0-4139-b309-4306411e6906 req-4f2e4a0e-1b87-4d1f-97e9-833d3d87e5a1 service nova] Acquiring lock "refresh_cache-f92fae48-7285-4b69-97c9-2360227c18fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.469475] env[62692]: DEBUG oslo_concurrency.lockutils [req-cc91433d-6ec0-4139-b309-4306411e6906 req-4f2e4a0e-1b87-4d1f-97e9-833d3d87e5a1 service nova] Acquired lock "refresh_cache-f92fae48-7285-4b69-97c9-2360227c18fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.469475] env[62692]: DEBUG nova.network.neutron [req-cc91433d-6ec0-4139-b309-4306411e6906 req-4f2e4a0e-1b87-4d1f-97e9-833d3d87e5a1 service nova] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Refreshing network info cache for port 3c66a7e2-2738-4873-b748-348237876eda {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1087.531419] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5f8ef3a9-5792-4e4b-923d-db31580910a5 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.233s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.545056] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141790, 'name': CreateVM_Task, 'duration_secs': 0.502855} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.545507] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1087.546195] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.546365] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.546666] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1087.546953] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ddb9ffeb-2ab8-4147-88e2-7feffad38f9c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.549414] env[62692]: DEBUG nova.scheduler.client.report [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1087.559705] env[62692]: DEBUG oslo_vmware.api [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 1087.559705] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52f3d22f-a07b-c4e4-26b0-a687d1eceed9" [ 1087.559705] env[62692]: _type = "Task" [ 1087.559705] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.567629] env[62692]: DEBUG oslo_vmware.api [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52f3d22f-a07b-c4e4-26b0-a687d1eceed9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.662253] env[62692]: DEBUG nova.network.neutron [req-f84cdc2f-cd43-4924-9993-9d1cad623ca0 req-c5b296e2-18bd-4dfd-a0ce-195ec28cfd21 service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Updated VIF entry in instance network info cache for port d51ffd3f-c023-45e2-8bfa-1424b7ccb188. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1087.662723] env[62692]: DEBUG nova.network.neutron [req-f84cdc2f-cd43-4924-9993-9d1cad623ca0 req-c5b296e2-18bd-4dfd-a0ce-195ec28cfd21 service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Updating instance_info_cache with network_info: [{"id": "d51ffd3f-c023-45e2-8bfa-1424b7ccb188", "address": "fa:16:3e:a8:44:f1", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd51ffd3f-c0", "ovs_interfaceid": "d51ffd3f-c023-45e2-8bfa-1424b7ccb188", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.668195] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "1904edb2-fd72-43b6-9219-8043c5abff0c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.668473] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "1904edb2-fd72-43b6-9219-8043c5abff0c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.686025] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141792, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.227834} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.686025] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1087.686230] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1087.687252] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1087.712718] env[62692]: DEBUG oslo_vmware.api [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Task: {'id': task-1141793, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069459} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.713017] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1087.713875] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52970265-2a17-4ff3-ac09-29b39d0de6e9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.738492] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] f27e3a4f-d109-472b-a348-5099bdc0e7c8/f27e3a4f-d109-472b-a348-5099bdc0e7c8.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1087.739476] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-188378f3-e7ea-4d02-8d68-1d99595969bc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.762202] env[62692]: DEBUG oslo_vmware.api [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Waiting for the task: (returnval){ [ 1087.762202] env[62692]: value = "task-1141794" [ 1087.762202] env[62692]: _type = "Task" [ 1087.762202] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.773871] env[62692]: DEBUG oslo_vmware.api [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Task: {'id': task-1141794, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.059352] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.394s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.059905] env[62692]: DEBUG nova.compute.manager [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1088.063599] env[62692]: DEBUG oslo_concurrency.lockutils [None req-84b3c9d5-bce5-416a-9561-94ae737dd0a5 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 4.278s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.074572] env[62692]: DEBUG oslo_vmware.api [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52f3d22f-a07b-c4e4-26b0-a687d1eceed9, 'name': SearchDatastore_Task, 'duration_secs': 0.029415} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.077781] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.077781] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1088.078009] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1088.078168] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1088.078387] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1088.079352] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-492badf3-99f3-431a-aa14-b884fdcc1524 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.089118] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1088.089778] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1088.090919] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88e3c9e2-3a48-4789-9783-ac62fd3438da {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.096643] env[62692]: DEBUG oslo_vmware.api [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 1088.096643] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52af0bc3-30f6-ea63-48e3-c39922e04632" [ 1088.096643] env[62692]: _type = "Task" [ 1088.096643] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.107972] env[62692]: DEBUG oslo_vmware.api [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52af0bc3-30f6-ea63-48e3-c39922e04632, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.168598] env[62692]: DEBUG oslo_concurrency.lockutils [req-f84cdc2f-cd43-4924-9993-9d1cad623ca0 req-c5b296e2-18bd-4dfd-a0ce-195ec28cfd21 service nova] Releasing lock "refresh_cache-94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.173400] env[62692]: DEBUG nova.compute.manager [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1088.192577] env[62692]: INFO nova.virt.block_device [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Booting with volume ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60 at /dev/sdb [ 1088.247227] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eb7ff340-a80f-4ffb-bf7d-d01b0ab30db9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.257948] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-293f12bb-191d-4b9e-a13b-40a169a258ed {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.281343] env[62692]: DEBUG oslo_vmware.api [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Task: {'id': task-1141794, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.294340] env[62692]: DEBUG nova.network.neutron [req-cc91433d-6ec0-4139-b309-4306411e6906 req-4f2e4a0e-1b87-4d1f-97e9-833d3d87e5a1 service nova] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Updated VIF entry in instance network info cache for port 3c66a7e2-2738-4873-b748-348237876eda. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1088.294714] env[62692]: DEBUG nova.network.neutron [req-cc91433d-6ec0-4139-b309-4306411e6906 req-4f2e4a0e-1b87-4d1f-97e9-833d3d87e5a1 service nova] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Updating instance_info_cache with network_info: [{"id": "3c66a7e2-2738-4873-b748-348237876eda", "address": "fa:16:3e:68:2e:27", "network": {"id": "d0e00428-1d04-4b63-8975-7f3085784358", "bridge": "br-int", "label": "tempest-ServersTestJSON-408035146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e13f8888d5e84eed8247101558e29a56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c66a7e2-27", "ovs_interfaceid": "3c66a7e2-2738-4873-b748-348237876eda", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.296100] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f0a9642a-5d42-46ad-9bbf-ce95f7917065 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.304409] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3554fdd6-96ad-4fa5-bc19-d4dd0a62024c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.335343] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52d95ab5-434d-442d-ae00-0fecb9e6acb6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.342386] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba3acc0d-2213-469c-aa96-143ee67b1833 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.356482] env[62692]: DEBUG nova.virt.block_device [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Updating existing volume attachment record: f18cf0b4-6979-4cfd-8a39-24093c55bfe9 {{(pid=62692) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1088.572826] env[62692]: DEBUG nova.compute.utils [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1088.575228] env[62692]: DEBUG nova.compute.manager [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1088.577302] env[62692]: DEBUG nova.network.neutron [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1088.609242] env[62692]: DEBUG oslo_vmware.api [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52af0bc3-30f6-ea63-48e3-c39922e04632, 'name': SearchDatastore_Task, 'duration_secs': 0.010399} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.612733] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8033e280-8e3d-45b0-9f98-ebcb1cf5aaa0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.617686] env[62692]: DEBUG oslo_vmware.api [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 1088.617686] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52780480-8a81-dfa1-752c-04953fbfed0d" [ 1088.617686] env[62692]: _type = "Task" [ 1088.617686] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.623395] env[62692]: DEBUG nova.policy [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a28f5e0e0af443c5be82214d446c221d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '589895ae20394e2fa525707bf7d09f2a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1088.630981] env[62692]: DEBUG oslo_vmware.api [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52780480-8a81-dfa1-752c-04953fbfed0d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.691800] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.783771] env[62692]: DEBUG oslo_vmware.api [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Task: {'id': task-1141794, 'name': ReconfigVM_Task, 'duration_secs': 0.685842} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.784286] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Reconfigured VM instance instance-00000063 to attach disk [datastore2] f27e3a4f-d109-472b-a348-5099bdc0e7c8/f27e3a4f-d109-472b-a348-5099bdc0e7c8.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1088.784997] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a7afa644-03c8-4f0b-a5d4-06bd98e2ff7a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.790445] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e6f8681-4a8b-4745-b9be-38269aa2a206 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.797172] env[62692]: DEBUG oslo_vmware.api [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Waiting for the task: (returnval){ [ 1088.797172] env[62692]: value = "task-1141795" [ 1088.797172] env[62692]: _type = "Task" [ 1088.797172] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.799439] env[62692]: DEBUG oslo_concurrency.lockutils [req-cc91433d-6ec0-4139-b309-4306411e6906 req-4f2e4a0e-1b87-4d1f-97e9-833d3d87e5a1 service nova] Releasing lock "refresh_cache-f92fae48-7285-4b69-97c9-2360227c18fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.803548] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41331c42-251d-4eab-8e46-e8c29b189e46 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.813058] env[62692]: DEBUG oslo_vmware.api [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Task: {'id': task-1141795, 'name': Rename_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.837345] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dfad9bc-552c-4776-ad4d-a22e91171d4d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.845033] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22226393-b070-4ea1-ab9e-4e5ef92a2ef1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.860659] env[62692]: DEBUG nova.compute.provider_tree [None req-84b3c9d5-bce5-416a-9561-94ae737dd0a5 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1088.949962] env[62692]: DEBUG nova.network.neutron [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Successfully created port: 65f2688b-0a0c-486b-a629-e9dc96ad98b6 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1089.080718] env[62692]: DEBUG nova.compute.manager [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1089.129871] env[62692]: DEBUG oslo_vmware.api [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52780480-8a81-dfa1-752c-04953fbfed0d, 'name': SearchDatastore_Task, 'duration_secs': 0.040867} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.130217] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.130506] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] f92fae48-7285-4b69-97c9-2360227c18fc/f92fae48-7285-4b69-97c9-2360227c18fc.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1089.130752] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dce8a8fb-8376-4883-826e-2d124806d437 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.137454] env[62692]: DEBUG oslo_vmware.api [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 1089.137454] env[62692]: value = "task-1141796" [ 1089.137454] env[62692]: _type = "Task" [ 1089.137454] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.146363] env[62692]: DEBUG oslo_vmware.api [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141796, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.310930] env[62692]: DEBUG oslo_vmware.api [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Task: {'id': task-1141795, 'name': Rename_Task, 'duration_secs': 0.199054} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.311307] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1089.311777] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f64a0525-166b-4eb9-bf2e-0e78a2d7801e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.319271] env[62692]: DEBUG oslo_vmware.api [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Waiting for the task: (returnval){ [ 1089.319271] env[62692]: value = "task-1141797" [ 1089.319271] env[62692]: _type = "Task" [ 1089.319271] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.330608] env[62692]: DEBUG oslo_vmware.api [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Task: {'id': task-1141797, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.364111] env[62692]: DEBUG nova.scheduler.client.report [None req-84b3c9d5-bce5-416a-9561-94ae737dd0a5 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1089.505438] env[62692]: DEBUG nova.compute.manager [req-383a86d6-301d-409d-b11a-2d7a9310b825 req-f6f17ac3-4a0b-440e-bd8a-8d2eeddaedaa service nova] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Received event network-changed-a3885c51-6f7b-41e8-8948-41fac442a288 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1089.506460] env[62692]: DEBUG nova.compute.manager [req-383a86d6-301d-409d-b11a-2d7a9310b825 req-f6f17ac3-4a0b-440e-bd8a-8d2eeddaedaa service nova] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Refreshing instance network info cache due to event network-changed-a3885c51-6f7b-41e8-8948-41fac442a288. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1089.506460] env[62692]: DEBUG oslo_concurrency.lockutils [req-383a86d6-301d-409d-b11a-2d7a9310b825 req-f6f17ac3-4a0b-440e-bd8a-8d2eeddaedaa service nova] Acquiring lock "refresh_cache-cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.506460] env[62692]: DEBUG oslo_concurrency.lockutils [req-383a86d6-301d-409d-b11a-2d7a9310b825 req-f6f17ac3-4a0b-440e-bd8a-8d2eeddaedaa service nova] Acquired lock "refresh_cache-cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.506460] env[62692]: DEBUG nova.network.neutron [req-383a86d6-301d-409d-b11a-2d7a9310b825 req-f6f17ac3-4a0b-440e-bd8a-8d2eeddaedaa service nova] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Refreshing network info cache for port a3885c51-6f7b-41e8-8948-41fac442a288 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1089.652073] env[62692]: DEBUG oslo_vmware.api [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141796, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.833941] env[62692]: DEBUG oslo_vmware.api [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Task: {'id': task-1141797, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.091065] env[62692]: DEBUG nova.compute.manager [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1090.118991] env[62692]: DEBUG nova.virt.hardware [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1090.119331] env[62692]: DEBUG nova.virt.hardware [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1090.119543] env[62692]: DEBUG nova.virt.hardware [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1090.119751] env[62692]: DEBUG nova.virt.hardware [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1090.119906] env[62692]: DEBUG nova.virt.hardware [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1090.120115] env[62692]: DEBUG nova.virt.hardware [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1090.120354] env[62692]: DEBUG nova.virt.hardware [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1090.120559] env[62692]: DEBUG nova.virt.hardware [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1090.120792] env[62692]: DEBUG nova.virt.hardware [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1090.120992] env[62692]: DEBUG nova.virt.hardware [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1090.121315] env[62692]: DEBUG nova.virt.hardware [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1090.122805] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2c19094-44ba-41b7-9f72-294628818875 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.130232] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a531319-8a29-49f0-9542-a16e37e2e522 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.153480] env[62692]: DEBUG oslo_vmware.api [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141796, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.812896} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.153741] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] f92fae48-7285-4b69-97c9-2360227c18fc/f92fae48-7285-4b69-97c9-2360227c18fc.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1090.153953] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1090.154222] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c920cff4-16ec-4349-a90a-e79d7a914e6b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.160707] env[62692]: DEBUG oslo_vmware.api [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 1090.160707] env[62692]: value = "task-1141798" [ 1090.160707] env[62692]: _type = "Task" [ 1090.160707] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.167309] env[62692]: DEBUG oslo_vmware.api [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141798, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.240671] env[62692]: DEBUG nova.network.neutron [req-383a86d6-301d-409d-b11a-2d7a9310b825 req-f6f17ac3-4a0b-440e-bd8a-8d2eeddaedaa service nova] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Updated VIF entry in instance network info cache for port a3885c51-6f7b-41e8-8948-41fac442a288. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1090.241150] env[62692]: DEBUG nova.network.neutron [req-383a86d6-301d-409d-b11a-2d7a9310b825 req-f6f17ac3-4a0b-440e-bd8a-8d2eeddaedaa service nova] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Updating instance_info_cache with network_info: [{"id": "a3885c51-6f7b-41e8-8948-41fac442a288", "address": "fa:16:3e:e4:9b:6f", "network": {"id": "f70727d6-a9aa-44b0-a244-2533b3f96571", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-284105837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1037e5d436bd429391159b87a712f6e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3885c51-6f", "ovs_interfaceid": "a3885c51-6f7b-41e8-8948-41fac442a288", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.331088] env[62692]: DEBUG oslo_vmware.api [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Task: {'id': task-1141797, 'name': PowerOnVM_Task, 'duration_secs': 0.606808} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.331088] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1090.331088] env[62692]: INFO nova.compute.manager [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Took 9.54 seconds to spawn the instance on the hypervisor. [ 1090.331088] env[62692]: DEBUG nova.compute.manager [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1090.332040] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73b31015-679a-43a3-8bf2-c674ff15d1d1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.377345] env[62692]: DEBUG oslo_concurrency.lockutils [None req-84b3c9d5-bce5-416a-9561-94ae737dd0a5 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.314s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.381076] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 5.653s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.479105] env[62692]: DEBUG nova.virt.hardware [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1090.479105] env[62692]: DEBUG nova.virt.hardware [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1090.479458] env[62692]: DEBUG nova.virt.hardware [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1090.479781] env[62692]: DEBUG nova.virt.hardware [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1090.480058] env[62692]: DEBUG nova.virt.hardware [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1090.480312] env[62692]: DEBUG nova.virt.hardware [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1090.480625] env[62692]: DEBUG nova.virt.hardware [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1090.480957] env[62692]: DEBUG nova.virt.hardware [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1090.481292] env[62692]: DEBUG nova.virt.hardware [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1090.481574] env[62692]: DEBUG nova.virt.hardware [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1090.481967] env[62692]: DEBUG nova.virt.hardware [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1090.482942] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57a6a512-59ac-473e-899b-3524588059d6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.492565] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca3cbdce-64f0-47f5-b324-885ea9999948 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.506797] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:73:e7:26', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '877ab637-ea08-499f-9d34-88ca15e1ed1a', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1090.515974] env[62692]: DEBUG oslo.service.loopingcall [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1090.516478] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1090.516905] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cfc0c3fd-f96c-4a35-a546-37d5940963d3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.539020] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1090.539020] env[62692]: value = "task-1141799" [ 1090.539020] env[62692]: _type = "Task" [ 1090.539020] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.544291] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141799, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.557293] env[62692]: DEBUG nova.network.neutron [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Successfully updated port: 65f2688b-0a0c-486b-a629-e9dc96ad98b6 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1090.670334] env[62692]: DEBUG oslo_vmware.api [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141798, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068102} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.670741] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1090.671389] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-555ea0b3-0dc1-4e38-a7d8-a69c7caf2979 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.694884] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] f92fae48-7285-4b69-97c9-2360227c18fc/f92fae48-7285-4b69-97c9-2360227c18fc.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1090.695227] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e84d3180-5ded-4ff3-b1ad-30fb37fa3b8b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.714661] env[62692]: DEBUG oslo_vmware.api [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 1090.714661] env[62692]: value = "task-1141800" [ 1090.714661] env[62692]: _type = "Task" [ 1090.714661] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.722994] env[62692]: DEBUG oslo_vmware.api [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141800, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.743896] env[62692]: DEBUG oslo_concurrency.lockutils [req-383a86d6-301d-409d-b11a-2d7a9310b825 req-f6f17ac3-4a0b-440e-bd8a-8d2eeddaedaa service nova] Releasing lock "refresh_cache-cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.848803] env[62692]: INFO nova.compute.manager [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Took 19.76 seconds to build instance. [ 1090.945118] env[62692]: INFO nova.scheduler.client.report [None req-84b3c9d5-bce5-416a-9561-94ae737dd0a5 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Deleted allocation for migration cfaf78e4-d48c-45bb-be40-e8e1c88b83f0 [ 1091.046679] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141799, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.059972] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "refresh_cache-1d9271aa-d41d-48cd-a879-3306e7212197" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1091.060303] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired lock "refresh_cache-1d9271aa-d41d-48cd-a879-3306e7212197" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.060303] env[62692]: DEBUG nova.network.neutron [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1091.225687] env[62692]: DEBUG oslo_vmware.api [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141800, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.350552] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3cf6ffc2-f118-4085-b60d-1c896bb15b1a tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Lock "f27e3a4f-d109-472b-a348-5099bdc0e7c8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.274s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.409869] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance e2383a6a-3581-40fc-a0eb-6981acdbf54a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1091.410050] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 7183ec4a-992d-48f1-8dda-7f499c2f4e1d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1091.410196] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance e8f645da-3b4a-4f21-9a17-aebdc8d112df actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1091.410334] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 90932f05-fad3-4c6d-87ae-ab059351b0be actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1091.410490] env[62692]: WARNING nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance e086d79d-29e6-41bf-a139-0b680ca0021d is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1091.410630] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1091.410794] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1091.410929] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 94a07505-661f-4595-9024-8a9da9f2c2fc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1091.411074] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1091.411202] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance f27e3a4f-d109-472b-a348-5099bdc0e7c8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1091.411351] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance f92fae48-7285-4b69-97c9-2360227c18fc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1091.411486] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 1d9271aa-d41d-48cd-a879-3306e7212197 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1091.451382] env[62692]: DEBUG oslo_concurrency.lockutils [None req-84b3c9d5-bce5-416a-9561-94ae737dd0a5 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 10.721s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.546505] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141799, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.583985] env[62692]: DEBUG nova.compute.manager [req-952b36e0-3216-4119-9735-82462a5486f9 req-e2655f8d-de82-4c3a-8f0a-86cb1895189b service nova] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Received event network-vif-plugged-65f2688b-0a0c-486b-a629-e9dc96ad98b6 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1091.584234] env[62692]: DEBUG oslo_concurrency.lockutils [req-952b36e0-3216-4119-9735-82462a5486f9 req-e2655f8d-de82-4c3a-8f0a-86cb1895189b service nova] Acquiring lock "1d9271aa-d41d-48cd-a879-3306e7212197-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.584445] env[62692]: DEBUG oslo_concurrency.lockutils [req-952b36e0-3216-4119-9735-82462a5486f9 req-e2655f8d-de82-4c3a-8f0a-86cb1895189b service nova] Lock "1d9271aa-d41d-48cd-a879-3306e7212197-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.584625] env[62692]: DEBUG oslo_concurrency.lockutils [req-952b36e0-3216-4119-9735-82462a5486f9 req-e2655f8d-de82-4c3a-8f0a-86cb1895189b service nova] Lock "1d9271aa-d41d-48cd-a879-3306e7212197-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.584843] env[62692]: DEBUG nova.compute.manager [req-952b36e0-3216-4119-9735-82462a5486f9 req-e2655f8d-de82-4c3a-8f0a-86cb1895189b service nova] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] No waiting events found dispatching network-vif-plugged-65f2688b-0a0c-486b-a629-e9dc96ad98b6 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1091.585015] env[62692]: WARNING nova.compute.manager [req-952b36e0-3216-4119-9735-82462a5486f9 req-e2655f8d-de82-4c3a-8f0a-86cb1895189b service nova] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Received unexpected event network-vif-plugged-65f2688b-0a0c-486b-a629-e9dc96ad98b6 for instance with vm_state building and task_state spawning. [ 1091.585192] env[62692]: DEBUG nova.compute.manager [req-952b36e0-3216-4119-9735-82462a5486f9 req-e2655f8d-de82-4c3a-8f0a-86cb1895189b service nova] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Received event network-changed-65f2688b-0a0c-486b-a629-e9dc96ad98b6 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1091.585355] env[62692]: DEBUG nova.compute.manager [req-952b36e0-3216-4119-9735-82462a5486f9 req-e2655f8d-de82-4c3a-8f0a-86cb1895189b service nova] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Refreshing instance network info cache due to event network-changed-65f2688b-0a0c-486b-a629-e9dc96ad98b6. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1091.585528] env[62692]: DEBUG oslo_concurrency.lockutils [req-952b36e0-3216-4119-9735-82462a5486f9 req-e2655f8d-de82-4c3a-8f0a-86cb1895189b service nova] Acquiring lock "refresh_cache-1d9271aa-d41d-48cd-a879-3306e7212197" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1091.594342] env[62692]: DEBUG nova.network.neutron [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1091.640847] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2652d64e-d810-44f4-9fb4-475eefa83baa tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "7183ec4a-992d-48f1-8dda-7f499c2f4e1d" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.641485] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2652d64e-d810-44f4-9fb4-475eefa83baa tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "7183ec4a-992d-48f1-8dda-7f499c2f4e1d" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.728128] env[62692]: DEBUG oslo_vmware.api [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141800, 'name': ReconfigVM_Task, 'duration_secs': 0.837078} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.728584] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Reconfigured VM instance instance-00000064 to attach disk [datastore2] f92fae48-7285-4b69-97c9-2360227c18fc/f92fae48-7285-4b69-97c9-2360227c18fc.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1091.729139] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b519f737-14ce-4a1c-a4dd-710ef1d2fdb3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.735414] env[62692]: DEBUG oslo_vmware.api [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 1091.735414] env[62692]: value = "task-1141801" [ 1091.735414] env[62692]: _type = "Task" [ 1091.735414] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.743474] env[62692]: DEBUG oslo_vmware.api [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141801, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.771511] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.771960] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.772312] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.772645] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.772890] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.775822] env[62692]: INFO nova.compute.manager [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Terminating instance [ 1091.777679] env[62692]: DEBUG nova.compute.manager [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1091.777882] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1091.778768] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ca5c0e9-4d5f-48c4-8147-e87bf8f4b879 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.787013] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1091.788038] env[62692]: DEBUG nova.network.neutron [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Updating instance_info_cache with network_info: [{"id": "65f2688b-0a0c-486b-a629-e9dc96ad98b6", "address": "fa:16:3e:96:e7:d8", "network": {"id": "dd260020-2d74-46ba-aa78-af29bdefa176", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-617179655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "589895ae20394e2fa525707bf7d09f2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65f2688b-0a", "ovs_interfaceid": "65f2688b-0a0c-486b-a629-e9dc96ad98b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.789130] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1979d5f0-efe3-4eca-b76e-71eb159bab7b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.799876] env[62692]: DEBUG oslo_vmware.api [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1091.799876] env[62692]: value = "task-1141802" [ 1091.799876] env[62692]: _type = "Task" [ 1091.799876] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.809898] env[62692]: DEBUG oslo_vmware.api [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141802, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.915218] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 1904edb2-fd72-43b6-9219-8043c5abff0c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1091.915578] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=62692) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1091.916050] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2688MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=62692) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1092.046441] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141799, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.095382] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f62afe1d-7c4d-489e-80ad-40c1caba3042 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.103181] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a64e9841-1240-4020-b4e8-6997c1d85031 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.134064] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83dc2efd-821c-4e7c-be71-a8e74a8a624d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.141379] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47de34bc-64b6-4b00-a4ac-7941ad2780fa {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.146032] env[62692]: INFO nova.compute.manager [None req-2652d64e-d810-44f4-9fb4-475eefa83baa tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Detaching volume fc2d0515-f41f-4b2d-932b-af3a09d72939 [ 1092.159630] env[62692]: DEBUG nova.compute.provider_tree [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1092.194737] env[62692]: INFO nova.virt.block_device [None req-2652d64e-d810-44f4-9fb4-475eefa83baa tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Attempting to driver detach volume fc2d0515-f41f-4b2d-932b-af3a09d72939 from mountpoint /dev/sdb [ 1092.195042] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-2652d64e-d810-44f4-9fb4-475eefa83baa tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Volume detach. Driver type: vmdk {{(pid=62692) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1092.195246] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-2652d64e-d810-44f4-9fb4-475eefa83baa tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249028', 'volume_id': 'fc2d0515-f41f-4b2d-932b-af3a09d72939', 'name': 'volume-fc2d0515-f41f-4b2d-932b-af3a09d72939', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7183ec4a-992d-48f1-8dda-7f499c2f4e1d', 'attached_at': '', 'detached_at': '', 'volume_id': 'fc2d0515-f41f-4b2d-932b-af3a09d72939', 'serial': 'fc2d0515-f41f-4b2d-932b-af3a09d72939'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1092.196150] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c77c7829-ac50-4436-a20b-ce2d4e7d305d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.219171] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f049027c-a167-45ff-a268-fb9ab6341e55 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.224712] env[62692]: DEBUG oslo_concurrency.lockutils [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "adfb0194-e94b-4a87-8554-a5c262fe9a70" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.225017] env[62692]: DEBUG oslo_concurrency.lockutils [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "adfb0194-e94b-4a87-8554-a5c262fe9a70" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.230158] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c30dab6e-baa4-4ce1-a653-40756342bb0e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.255210] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-747c26b3-c6ec-4382-8267-4c53f6982219 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.262697] env[62692]: DEBUG oslo_vmware.api [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141801, 'name': Rename_Task, 'duration_secs': 0.153382} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.273013] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1092.273765] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-2652d64e-d810-44f4-9fb4-475eefa83baa tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] The volume has not been displaced from its original location: [datastore2] volume-fc2d0515-f41f-4b2d-932b-af3a09d72939/volume-fc2d0515-f41f-4b2d-932b-af3a09d72939.vmdk. No consolidation needed. {{(pid=62692) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1092.278945] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-2652d64e-d810-44f4-9fb4-475eefa83baa tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Reconfiguring VM instance instance-00000055 to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1092.279269] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-964576b0-8c8d-49f5-be79-02193ff329ee {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.280823] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b744d10-4cad-4b83-a3bf-a54e4351fe61 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.294099] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Releasing lock "refresh_cache-1d9271aa-d41d-48cd-a879-3306e7212197" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1092.294393] env[62692]: DEBUG nova.compute.manager [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Instance network_info: |[{"id": "65f2688b-0a0c-486b-a629-e9dc96ad98b6", "address": "fa:16:3e:96:e7:d8", "network": {"id": "dd260020-2d74-46ba-aa78-af29bdefa176", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-617179655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "589895ae20394e2fa525707bf7d09f2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65f2688b-0a", "ovs_interfaceid": "65f2688b-0a0c-486b-a629-e9dc96ad98b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1092.295124] env[62692]: DEBUG oslo_concurrency.lockutils [req-952b36e0-3216-4119-9735-82462a5486f9 req-e2655f8d-de82-4c3a-8f0a-86cb1895189b service nova] Acquired lock "refresh_cache-1d9271aa-d41d-48cd-a879-3306e7212197" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.295343] env[62692]: DEBUG nova.network.neutron [req-952b36e0-3216-4119-9735-82462a5486f9 req-e2655f8d-de82-4c3a-8f0a-86cb1895189b service nova] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Refreshing network info cache for port 65f2688b-0a0c-486b-a629-e9dc96ad98b6 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1092.296492] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:96:e7:d8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55764410-260e-4339-a020-6b30995584bf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '65f2688b-0a0c-486b-a629-e9dc96ad98b6', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1092.303548] env[62692]: DEBUG oslo.service.loopingcall [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1092.306041] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1092.306783] env[62692]: DEBUG oslo_vmware.api [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 1092.306783] env[62692]: value = "task-1141803" [ 1092.306783] env[62692]: _type = "Task" [ 1092.306783] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.310031] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1392481f-1fee-4629-a682-fd492019d1ed {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.334356] env[62692]: DEBUG oslo_vmware.api [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141802, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.361658] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1092.361658] env[62692]: value = "task-1141805" [ 1092.361658] env[62692]: _type = "Task" [ 1092.361658] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.362292] env[62692]: DEBUG oslo_vmware.api [None req-2652d64e-d810-44f4-9fb4-475eefa83baa tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1092.362292] env[62692]: value = "task-1141804" [ 1092.362292] env[62692]: _type = "Task" [ 1092.362292] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.377263] env[62692]: DEBUG oslo_vmware.api [None req-2652d64e-d810-44f4-9fb4-475eefa83baa tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141804, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.377805] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141805, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.549083] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141799, 'name': CreateVM_Task, 'duration_secs': 1.587725} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.549083] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1092.549307] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.549375] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.549739] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1092.550749] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ccafafa-cc8d-424a-bb5a-5c0f77439ee0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.556370] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1092.556370] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]527a0741-fb5c-af96-11e7-47a77aa5ffc4" [ 1092.556370] env[62692]: _type = "Task" [ 1092.556370] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.565742] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]527a0741-fb5c-af96-11e7-47a77aa5ffc4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.664497] env[62692]: DEBUG nova.scheduler.client.report [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1092.727881] env[62692]: DEBUG nova.compute.manager [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1092.817031] env[62692]: DEBUG oslo_vmware.api [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141802, 'name': PowerOffVM_Task, 'duration_secs': 0.600129} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.817387] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1092.817425] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1092.817661] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0584dbb4-5992-4a65-bd02-c2d21b8e1363 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.835300] env[62692]: DEBUG oslo_vmware.api [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141803, 'name': PowerOnVM_Task, 'duration_secs': 0.522487} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.835579] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1092.835784] env[62692]: INFO nova.compute.manager [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Took 7.98 seconds to spawn the instance on the hypervisor. [ 1092.835971] env[62692]: DEBUG nova.compute.manager [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1092.836752] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8389c5a1-37d2-48a7-b9f5-70feabae0887 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.849259] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Acquiring lock "f27e3a4f-d109-472b-a348-5099bdc0e7c8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.849549] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Lock "f27e3a4f-d109-472b-a348-5099bdc0e7c8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.849775] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Acquiring lock "f27e3a4f-d109-472b-a348-5099bdc0e7c8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.850257] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Lock "f27e3a4f-d109-472b-a348-5099bdc0e7c8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.850461] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Lock "f27e3a4f-d109-472b-a348-5099bdc0e7c8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.853181] env[62692]: INFO nova.compute.manager [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Terminating instance [ 1092.854765] env[62692]: DEBUG nova.compute.manager [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1092.854967] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1092.855801] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33743cfd-606d-4f92-b69e-689ec09d6bc7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.863582] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1092.870031] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-85c1e9bf-8c4e-424d-916e-b621f7f7df87 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.876366] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141805, 'name': CreateVM_Task, 'duration_secs': 0.35348} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.880606] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1092.881037] env[62692]: DEBUG oslo_vmware.api [None req-2652d64e-d810-44f4-9fb4-475eefa83baa tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141804, 'name': ReconfigVM_Task, 'duration_secs': 0.357392} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.881356] env[62692]: DEBUG oslo_vmware.api [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Waiting for the task: (returnval){ [ 1092.881356] env[62692]: value = "task-1141807" [ 1092.881356] env[62692]: _type = "Task" [ 1092.881356] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.882866] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.883143] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-2652d64e-d810-44f4-9fb4-475eefa83baa tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Reconfigured VM instance instance-00000055 to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1092.888048] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1092.888271] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1092.888433] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Deleting the datastore file [datastore2] 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1092.889440] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-27d1df88-a861-4c52-a9b5-81b728971db8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.904230] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3624f2ae-55b1-4d52-9f8f-6b1ab06e849e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.913263] env[62692]: DEBUG oslo_vmware.api [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Task: {'id': task-1141807, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.915692] env[62692]: DEBUG oslo_vmware.api [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1092.915692] env[62692]: value = "task-1141809" [ 1092.915692] env[62692]: _type = "Task" [ 1092.915692] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.915853] env[62692]: DEBUG oslo_vmware.api [None req-2652d64e-d810-44f4-9fb4-475eefa83baa tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1092.915853] env[62692]: value = "task-1141808" [ 1092.915853] env[62692]: _type = "Task" [ 1092.915853] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.926307] env[62692]: DEBUG oslo_vmware.api [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141809, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.929402] env[62692]: DEBUG oslo_vmware.api [None req-2652d64e-d810-44f4-9fb4-475eefa83baa tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141808, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.067143] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]527a0741-fb5c-af96-11e7-47a77aa5ffc4, 'name': SearchDatastore_Task, 'duration_secs': 0.01536} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.067428] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.067681] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1093.067966] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1093.068115] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.068339] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1093.068665] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.069051] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1093.069317] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-44f263c2-7b28-4dd9-9f19-eb577254c36e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.071339] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f01a376-af6f-4d42-ac89-e7897dd10ddb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.080910] env[62692]: DEBUG oslo_vmware.api [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1093.080910] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52e35d1f-77e2-2c09-154b-37276b0304a3" [ 1093.080910] env[62692]: _type = "Task" [ 1093.080910] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.090661] env[62692]: DEBUG oslo_vmware.api [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52e35d1f-77e2-2c09-154b-37276b0304a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.091918] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1093.092112] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1093.095269] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b7f3a53-367d-48d6-b347-0f657f25e7f8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.101851] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1093.101851] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5260f62a-206d-3349-a25f-ecea3cb880d6" [ 1093.101851] env[62692]: _type = "Task" [ 1093.101851] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.109405] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5260f62a-206d-3349-a25f-ecea3cb880d6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.133676] env[62692]: DEBUG nova.network.neutron [req-952b36e0-3216-4119-9735-82462a5486f9 req-e2655f8d-de82-4c3a-8f0a-86cb1895189b service nova] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Updated VIF entry in instance network info cache for port 65f2688b-0a0c-486b-a629-e9dc96ad98b6. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1093.134044] env[62692]: DEBUG nova.network.neutron [req-952b36e0-3216-4119-9735-82462a5486f9 req-e2655f8d-de82-4c3a-8f0a-86cb1895189b service nova] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Updating instance_info_cache with network_info: [{"id": "65f2688b-0a0c-486b-a629-e9dc96ad98b6", "address": "fa:16:3e:96:e7:d8", "network": {"id": "dd260020-2d74-46ba-aa78-af29bdefa176", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-617179655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "589895ae20394e2fa525707bf7d09f2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65f2688b-0a", "ovs_interfaceid": "65f2688b-0a0c-486b-a629-e9dc96ad98b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.170277] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62692) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1093.170485] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.790s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.170835] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.589s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.171075] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.173352] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.482s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.174890] env[62692]: INFO nova.compute.claims [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1093.202200] env[62692]: INFO nova.scheduler.client.report [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Deleted allocations for instance e086d79d-29e6-41bf-a139-0b680ca0021d [ 1093.248209] env[62692]: DEBUG oslo_concurrency.lockutils [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.360895] env[62692]: INFO nova.compute.manager [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Took 18.68 seconds to build instance. [ 1093.397424] env[62692]: DEBUG oslo_vmware.api [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Task: {'id': task-1141807, 'name': PowerOffVM_Task, 'duration_secs': 0.186312} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.397696] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1093.397869] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1093.398129] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a57e7912-de32-4aec-94ae-d0f70912a032 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.427701] env[62692]: DEBUG oslo_vmware.api [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141809, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.181214} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.430755] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1093.431023] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1093.431227] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1093.431438] env[62692]: INFO nova.compute.manager [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Took 1.65 seconds to destroy the instance on the hypervisor. [ 1093.431701] env[62692]: DEBUG oslo.service.loopingcall [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1093.431951] env[62692]: DEBUG oslo_vmware.api [None req-2652d64e-d810-44f4-9fb4-475eefa83baa tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141808, 'name': ReconfigVM_Task, 'duration_secs': 0.185018} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.432208] env[62692]: DEBUG nova.compute.manager [-] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1093.432308] env[62692]: DEBUG nova.network.neutron [-] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1093.433977] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-2652d64e-d810-44f4-9fb4-475eefa83baa tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249028', 'volume_id': 'fc2d0515-f41f-4b2d-932b-af3a09d72939', 'name': 'volume-fc2d0515-f41f-4b2d-932b-af3a09d72939', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7183ec4a-992d-48f1-8dda-7f499c2f4e1d', 'attached_at': '', 'detached_at': '', 'volume_id': 'fc2d0515-f41f-4b2d-932b-af3a09d72939', 'serial': 'fc2d0515-f41f-4b2d-932b-af3a09d72939'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1093.465104] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1093.465357] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1093.465549] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Deleting the datastore file [datastore2] f27e3a4f-d109-472b-a348-5099bdc0e7c8 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1093.465827] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a09e622e-bafa-471a-bf29-3e99f6dd6e6f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.472246] env[62692]: DEBUG oslo_vmware.api [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Waiting for the task: (returnval){ [ 1093.472246] env[62692]: value = "task-1141811" [ 1093.472246] env[62692]: _type = "Task" [ 1093.472246] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.480438] env[62692]: DEBUG oslo_vmware.api [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Task: {'id': task-1141811, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.592142] env[62692]: DEBUG oslo_vmware.api [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52e35d1f-77e2-2c09-154b-37276b0304a3, 'name': SearchDatastore_Task, 'duration_secs': 0.025653} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.596519] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.596519] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1093.596519] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1093.613455] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5260f62a-206d-3349-a25f-ecea3cb880d6, 'name': SearchDatastore_Task, 'duration_secs': 0.009726} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.614185] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eaffac5d-89d8-45d7-8333-b0ccf4d0ddf4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.619501] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1093.619501] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52fd1996-7694-3ee8-2f06-111584ef4254" [ 1093.619501] env[62692]: _type = "Task" [ 1093.619501] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.629643] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52fd1996-7694-3ee8-2f06-111584ef4254, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.636812] env[62692]: DEBUG oslo_concurrency.lockutils [req-952b36e0-3216-4119-9735-82462a5486f9 req-e2655f8d-de82-4c3a-8f0a-86cb1895189b service nova] Releasing lock "refresh_cache-1d9271aa-d41d-48cd-a879-3306e7212197" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.710332] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e868cbb3-502a-4b3a-822c-1ba9f8bf7d26 tempest-ServerShowV254Test-393335733 tempest-ServerShowV254Test-393335733-project-member] Lock "e086d79d-29e6-41bf-a139-0b680ca0021d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.860s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.809257] env[62692]: DEBUG nova.compute.manager [req-2d98dc30-5018-4da7-8019-b092ac6c84d9 req-d8ed4b0f-ce03-4099-be5a-03247ffed3d3 service nova] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Received event network-vif-deleted-d5ff0422-e9b3-455c-bd28-422096794288 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1093.809476] env[62692]: INFO nova.compute.manager [req-2d98dc30-5018-4da7-8019-b092ac6c84d9 req-d8ed4b0f-ce03-4099-be5a-03247ffed3d3 service nova] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Neutron deleted interface d5ff0422-e9b3-455c-bd28-422096794288; detaching it from the instance and deleting it from the info cache [ 1093.809656] env[62692]: DEBUG nova.network.neutron [req-2d98dc30-5018-4da7-8019-b092ac6c84d9 req-d8ed4b0f-ce03-4099-be5a-03247ffed3d3 service nova] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.863084] env[62692]: DEBUG oslo_concurrency.lockutils [None req-28db5857-6931-4b4b-875a-6d43c3f0b34c tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "f92fae48-7285-4b69-97c9-2360227c18fc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.194s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.981068] env[62692]: DEBUG nova.objects.instance [None req-2652d64e-d810-44f4-9fb4-475eefa83baa tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lazy-loading 'flavor' on Instance uuid 7183ec4a-992d-48f1-8dda-7f499c2f4e1d {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1093.985421] env[62692]: DEBUG oslo_vmware.api [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Task: {'id': task-1141811, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.366577} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.985556] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1093.985763] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1093.985941] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1093.986300] env[62692]: INFO nova.compute.manager [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1093.986562] env[62692]: DEBUG oslo.service.loopingcall [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1093.986915] env[62692]: DEBUG nova.compute.manager [-] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1093.986915] env[62692]: DEBUG nova.network.neutron [-] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1094.130216] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52fd1996-7694-3ee8-2f06-111584ef4254, 'name': SearchDatastore_Task, 'duration_secs': 0.021304} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.130554] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1094.130847] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] e8f645da-3b4a-4f21-9a17-aebdc8d112df/e8f645da-3b4a-4f21-9a17-aebdc8d112df.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1094.131148] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.131344] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1094.131561] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-63e2e5a1-9682-42ac-9566-84545a9bb8ba {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.133528] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-86ee5e0f-cfac-4533-83fd-ca843a4421fb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.141306] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1094.141306] env[62692]: value = "task-1141812" [ 1094.141306] env[62692]: _type = "Task" [ 1094.141306] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.142413] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1094.142579] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1094.145941] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74691fa4-cebb-466d-8d54-c2d098971780 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.151189] env[62692]: DEBUG oslo_vmware.api [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1094.151189] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52e06722-b444-c28c-edc6-62ec7d40c6f6" [ 1094.151189] env[62692]: _type = "Task" [ 1094.151189] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.153874] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141812, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.161040] env[62692]: DEBUG oslo_vmware.api [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52e06722-b444-c28c-edc6-62ec7d40c6f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.242413] env[62692]: DEBUG nova.network.neutron [-] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.313339] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-23df5192-572a-4631-98d7-a496e783a65e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.331015] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d6d03b2-9941-4797-b1c3-e9dae0dc8bda {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.376074] env[62692]: DEBUG nova.compute.manager [req-2d98dc30-5018-4da7-8019-b092ac6c84d9 req-d8ed4b0f-ce03-4099-be5a-03247ffed3d3 service nova] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Detach interface failed, port_id=d5ff0422-e9b3-455c-bd28-422096794288, reason: Instance 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1094.422562] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46d8219e-7c54-4d1d-a475-3f44d2587660 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.430547] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8571ee54-be8e-4720-8bf0-0254f75bd77e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.465868] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-153dc6b1-d45c-45d0-8abc-cddbc4164cc0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.474273] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8ea6cfc-8a53-4a9b-8c61-5654addf6192 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.492481] env[62692]: DEBUG nova.compute.provider_tree [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1094.652368] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141812, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.663175] env[62692]: DEBUG oslo_vmware.api [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52e06722-b444-c28c-edc6-62ec7d40c6f6, 'name': SearchDatastore_Task, 'duration_secs': 0.021159} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.664130] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0417ebf5-792d-40cf-ae4d-44f8b5f3830e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.669607] env[62692]: DEBUG oslo_vmware.api [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1094.669607] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]520bfbb9-00c2-4126-808a-d65091ef17e3" [ 1094.669607] env[62692]: _type = "Task" [ 1094.669607] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.677093] env[62692]: DEBUG oslo_vmware.api [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]520bfbb9-00c2-4126-808a-d65091ef17e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.746354] env[62692]: INFO nova.compute.manager [-] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Took 1.31 seconds to deallocate network for instance. [ 1094.853388] env[62692]: DEBUG oslo_concurrency.lockutils [None req-83b7dcdc-d22b-4b6c-a381-836e8b64aa60 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "7183ec4a-992d-48f1-8dda-7f499c2f4e1d" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1094.996277] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2652d64e-d810-44f4-9fb4-475eefa83baa tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "7183ec4a-992d-48f1-8dda-7f499c2f4e1d" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.355s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.998467] env[62692]: DEBUG nova.scheduler.client.report [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1095.002686] env[62692]: DEBUG oslo_concurrency.lockutils [None req-83b7dcdc-d22b-4b6c-a381-836e8b64aa60 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "7183ec4a-992d-48f1-8dda-7f499c2f4e1d" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.150s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.003069] env[62692]: DEBUG nova.compute.manager [None req-83b7dcdc-d22b-4b6c-a381-836e8b64aa60 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1095.004327] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed55ebd1-78d9-449e-904e-76ff0eaf400e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.013820] env[62692]: DEBUG nova.compute.manager [None req-83b7dcdc-d22b-4b6c-a381-836e8b64aa60 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62692) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1095.014740] env[62692]: DEBUG nova.objects.instance [None req-83b7dcdc-d22b-4b6c-a381-836e8b64aa60 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lazy-loading 'flavor' on Instance uuid 7183ec4a-992d-48f1-8dda-7f499c2f4e1d {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1095.057152] env[62692]: DEBUG nova.network.neutron [-] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.152560] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141812, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.931515} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.152793] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] e8f645da-3b4a-4f21-9a17-aebdc8d112df/e8f645da-3b4a-4f21-9a17-aebdc8d112df.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1095.153022] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1095.153265] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1ab61a75-da25-4684-9ece-5a86234362a2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.159189] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1095.159189] env[62692]: value = "task-1141813" [ 1095.159189] env[62692]: _type = "Task" [ 1095.159189] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.166692] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141813, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.179757] env[62692]: DEBUG oslo_vmware.api [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]520bfbb9-00c2-4126-808a-d65091ef17e3, 'name': SearchDatastore_Task, 'duration_secs': 0.011309} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.180020] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.180293] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 1d9271aa-d41d-48cd-a879-3306e7212197/1d9271aa-d41d-48cd-a879-3306e7212197.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1095.180551] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ae29e6c9-ef06-4ecd-86a6-1e13a42b21c4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.185794] env[62692]: DEBUG oslo_vmware.api [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1095.185794] env[62692]: value = "task-1141814" [ 1095.185794] env[62692]: _type = "Task" [ 1095.185794] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.192910] env[62692]: DEBUG oslo_vmware.api [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141814, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.243143] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2b903274-b828-4192-b543-d1abb4f3e664 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "f92fae48-7285-4b69-97c9-2360227c18fc" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.243426] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2b903274-b828-4192-b543-d1abb4f3e664 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "f92fae48-7285-4b69-97c9-2360227c18fc" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.243615] env[62692]: DEBUG nova.compute.manager [None req-2b903274-b828-4192-b543-d1abb4f3e664 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1095.244618] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a0e827f-b990-4e5d-89e2-abb47f7054b1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.251170] env[62692]: DEBUG nova.compute.manager [None req-2b903274-b828-4192-b543-d1abb4f3e664 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62692) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1095.251755] env[62692]: DEBUG nova.objects.instance [None req-2b903274-b828-4192-b543-d1abb4f3e664 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lazy-loading 'flavor' on Instance uuid f92fae48-7285-4b69-97c9-2360227c18fc {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1095.253935] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.506078] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.332s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.506078] env[62692]: DEBUG nova.compute.manager [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1095.508884] env[62692]: DEBUG oslo_concurrency.lockutils [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.261s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.510375] env[62692]: INFO nova.compute.claims [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1095.519896] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-83b7dcdc-d22b-4b6c-a381-836e8b64aa60 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1095.520161] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8fd493fb-24e0-4cb4-89c8-195dfb44e17e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.526593] env[62692]: DEBUG oslo_vmware.api [None req-83b7dcdc-d22b-4b6c-a381-836e8b64aa60 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1095.526593] env[62692]: value = "task-1141815" [ 1095.526593] env[62692]: _type = "Task" [ 1095.526593] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.535116] env[62692]: DEBUG oslo_vmware.api [None req-83b7dcdc-d22b-4b6c-a381-836e8b64aa60 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141815, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.561148] env[62692]: INFO nova.compute.manager [-] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Took 1.57 seconds to deallocate network for instance. [ 1095.670571] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141813, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.289995} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.670847] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1095.671649] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11c69e43-19dc-43e7-acaf-78017ad72fb7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.695767] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] e8f645da-3b4a-4f21-9a17-aebdc8d112df/e8f645da-3b4a-4f21-9a17-aebdc8d112df.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1095.699380] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06940579-62d5-4946-aff1-d9171b4c95f3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.720090] env[62692]: DEBUG oslo_vmware.api [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141814, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.721673] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1095.721673] env[62692]: value = "task-1141816" [ 1095.721673] env[62692]: _type = "Task" [ 1095.721673] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.730518] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141816, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.758223] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b903274-b828-4192-b543-d1abb4f3e664 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1095.758740] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5caef905-07b7-4e28-9e11-423f92d7c705 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.766019] env[62692]: DEBUG oslo_vmware.api [None req-2b903274-b828-4192-b543-d1abb4f3e664 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 1095.766019] env[62692]: value = "task-1141817" [ 1095.766019] env[62692]: _type = "Task" [ 1095.766019] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.775175] env[62692]: DEBUG oslo_vmware.api [None req-2b903274-b828-4192-b543-d1abb4f3e664 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141817, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.839600] env[62692]: DEBUG nova.compute.manager [req-791da064-5b20-45c1-a01b-321e99241354 req-9526fa00-01fd-4a2e-8cf7-3c0485857fc6 service nova] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Received event network-vif-deleted-06a8d751-3ffa-4fb1-8017-a12d81fc1d98 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1096.014817] env[62692]: DEBUG nova.compute.utils [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1096.018842] env[62692]: DEBUG nova.compute.manager [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1096.019104] env[62692]: DEBUG nova.network.neutron [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1096.039355] env[62692]: DEBUG oslo_vmware.api [None req-83b7dcdc-d22b-4b6c-a381-836e8b64aa60 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141815, 'name': PowerOffVM_Task, 'duration_secs': 0.33859} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.039716] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-83b7dcdc-d22b-4b6c-a381-836e8b64aa60 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1096.040069] env[62692]: DEBUG nova.compute.manager [None req-83b7dcdc-d22b-4b6c-a381-836e8b64aa60 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1096.041030] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fec9686-4ec8-4368-9b36-26d0a4a32063 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.065780] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.084225] env[62692]: DEBUG nova.policy [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a251ef5f6cb143178928babc75a0fef0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ba97a775e99a463c845518b0df8cd3e2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1096.197970] env[62692]: DEBUG oslo_vmware.api [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141814, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.733156} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.198497] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 1d9271aa-d41d-48cd-a879-3306e7212197/1d9271aa-d41d-48cd-a879-3306e7212197.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1096.198815] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1096.199098] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2c7cce3e-d567-44dc-98e9-66015eb77b92 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.205379] env[62692]: DEBUG oslo_vmware.api [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1096.205379] env[62692]: value = "task-1141818" [ 1096.205379] env[62692]: _type = "Task" [ 1096.205379] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.214089] env[62692]: DEBUG oslo_vmware.api [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141818, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.233076] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141816, 'name': ReconfigVM_Task, 'duration_secs': 0.45068} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.233406] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Reconfigured VM instance instance-00000059 to attach disk [datastore1] e8f645da-3b4a-4f21-9a17-aebdc8d112df/e8f645da-3b4a-4f21-9a17-aebdc8d112df.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1096.234878] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'encryption_format': None, 'encryption_secret_uuid': None, 'guest_format': None, 'device_type': 'disk', 'device_name': '/dev/sda', 'size': 0, 'encrypted': False, 'encryption_options': None, 'disk_bus': None, 'boot_index': 0, 'image_id': '81eb64fb-5a0c-47b7-9948-7dcb01b911bd'}], 'ephemerals': [], 'block_device_mapping': [{'mount_device': '/dev/sdb', 'attachment_id': 'f18cf0b4-6979-4cfd-8a39-24093c55bfe9', 'guest_format': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249042', 'volume_id': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'name': 'volume-ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e8f645da-3b4a-4f21-9a17-aebdc8d112df', 'attached_at': '', 'detached_at': '', 'volume_id': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'serial': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60'}, 'device_type': None, 'delete_on_termination': False, 'disk_bus': None, 'boot_index': None, 'volume_type': None}], 'swap': None} {{(pid=62692) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1096.235140] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Volume attach. Driver type: vmdk {{(pid=62692) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1096.235367] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249042', 'volume_id': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'name': 'volume-ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e8f645da-3b4a-4f21-9a17-aebdc8d112df', 'attached_at': '', 'detached_at': '', 'volume_id': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'serial': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1096.236183] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26664f9d-8e87-422d-9f08-44fb83f4831a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.254922] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d953675b-a340-45f1-824c-f056e1b679dd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.280737] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] volume-ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60/volume-ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1096.283863] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a074a9b8-6d01-4985-be78-8dd4e83fa571 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.301755] env[62692]: DEBUG oslo_vmware.api [None req-2b903274-b828-4192-b543-d1abb4f3e664 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141817, 'name': PowerOffVM_Task, 'duration_secs': 0.282881} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.302951] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b903274-b828-4192-b543-d1abb4f3e664 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1096.303202] env[62692]: DEBUG nova.compute.manager [None req-2b903274-b828-4192-b543-d1abb4f3e664 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1096.303529] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1096.303529] env[62692]: value = "task-1141819" [ 1096.303529] env[62692]: _type = "Task" [ 1096.303529] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.304246] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-544c13f3-ef43-4415-aef2-d564ff39a211 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.314075] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141819, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.389861] env[62692]: DEBUG nova.network.neutron [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Successfully created port: a668d6b6-4111-4a94-8bc8-09535b29f895 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1096.523960] env[62692]: DEBUG nova.compute.manager [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1096.554060] env[62692]: DEBUG oslo_concurrency.lockutils [None req-83b7dcdc-d22b-4b6c-a381-836e8b64aa60 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "7183ec4a-992d-48f1-8dda-7f499c2f4e1d" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.551s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.707319] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd8e3a74-69d2-4c0c-9137-b023303ae7dc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.719989] env[62692]: DEBUG oslo_vmware.api [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141818, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067529} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.720971] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3bee552-44c5-48a2-b551-8f9a7c285047 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.723904] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1096.724626] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f92bca2-ecab-4199-8654-42c18f031a3c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.747200] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 1d9271aa-d41d-48cd-a879-3306e7212197/1d9271aa-d41d-48cd-a879-3306e7212197.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1096.773358] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af3f0545-2a7a-4568-ad6d-ea2cd60a42af {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.788058] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd2baff-de04-4f9d-8278-543cbfea3939 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.796769] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e26e2a3-3867-4fe2-8f1d-edbde57ce32c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.800535] env[62692]: DEBUG oslo_vmware.api [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1096.800535] env[62692]: value = "task-1141820" [ 1096.800535] env[62692]: _type = "Task" [ 1096.800535] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.815029] env[62692]: DEBUG nova.compute.provider_tree [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1096.820860] env[62692]: DEBUG oslo_vmware.api [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141820, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.822366] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2b903274-b828-4192-b543-d1abb4f3e664 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "f92fae48-7285-4b69-97c9-2360227c18fc" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.579s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.828741] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141819, 'name': ReconfigVM_Task, 'duration_secs': 0.314287} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.829608] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Reconfigured VM instance instance-00000059 to attach disk [datastore2] volume-ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60/volume-ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1096.834778] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9bff22eb-ac5e-441c-8c25-9ef2f4325d52 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.851587] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1096.851587] env[62692]: value = "task-1141821" [ 1096.851587] env[62692]: _type = "Task" [ 1096.851587] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.861138] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141821, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.310606] env[62692]: DEBUG oslo_vmware.api [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141820, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.317194] env[62692]: DEBUG nova.scheduler.client.report [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1097.331029] env[62692]: DEBUG nova.objects.instance [None req-cb7ff3cb-3a9d-4d8e-907e-4676d6ecc8ae tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lazy-loading 'flavor' on Instance uuid 7183ec4a-992d-48f1-8dda-7f499c2f4e1d {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1097.362631] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141821, 'name': ReconfigVM_Task, 'duration_secs': 0.240781} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.362631] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249042', 'volume_id': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'name': 'volume-ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e8f645da-3b4a-4f21-9a17-aebdc8d112df', 'attached_at': '', 'detached_at': '', 'volume_id': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'serial': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1097.362631] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-009c3914-8a51-4d3c-b907-3bac152e1757 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.369824] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1097.369824] env[62692]: value = "task-1141822" [ 1097.369824] env[62692]: _type = "Task" [ 1097.369824] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.378630] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141822, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.538465] env[62692]: DEBUG nova.compute.manager [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1097.562626] env[62692]: DEBUG nova.virt.hardware [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1097.562803] env[62692]: DEBUG nova.virt.hardware [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1097.562975] env[62692]: DEBUG nova.virt.hardware [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1097.563207] env[62692]: DEBUG nova.virt.hardware [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1097.563377] env[62692]: DEBUG nova.virt.hardware [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1097.563536] env[62692]: DEBUG nova.virt.hardware [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1097.563752] env[62692]: DEBUG nova.virt.hardware [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1097.563919] env[62692]: DEBUG nova.virt.hardware [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1097.564107] env[62692]: DEBUG nova.virt.hardware [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1097.564283] env[62692]: DEBUG nova.virt.hardware [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1097.564464] env[62692]: DEBUG nova.virt.hardware [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1097.565680] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e33ecc7-6963-41f3-99ea-f6e859bd6a96 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.574457] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bcc9181-2bcc-4329-9e8e-d035cd9f00e9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.756945] env[62692]: DEBUG nova.compute.manager [req-4fafd5ce-576e-481f-b9f6-640da3cf5bf1 req-791b0683-6c21-4ab6-972e-2c3074de3eb9 service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Received event network-vif-plugged-a668d6b6-4111-4a94-8bc8-09535b29f895 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1097.757198] env[62692]: DEBUG oslo_concurrency.lockutils [req-4fafd5ce-576e-481f-b9f6-640da3cf5bf1 req-791b0683-6c21-4ab6-972e-2c3074de3eb9 service nova] Acquiring lock "1904edb2-fd72-43b6-9219-8043c5abff0c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.757453] env[62692]: DEBUG oslo_concurrency.lockutils [req-4fafd5ce-576e-481f-b9f6-640da3cf5bf1 req-791b0683-6c21-4ab6-972e-2c3074de3eb9 service nova] Lock "1904edb2-fd72-43b6-9219-8043c5abff0c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.757583] env[62692]: DEBUG oslo_concurrency.lockutils [req-4fafd5ce-576e-481f-b9f6-640da3cf5bf1 req-791b0683-6c21-4ab6-972e-2c3074de3eb9 service nova] Lock "1904edb2-fd72-43b6-9219-8043c5abff0c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.757768] env[62692]: DEBUG nova.compute.manager [req-4fafd5ce-576e-481f-b9f6-640da3cf5bf1 req-791b0683-6c21-4ab6-972e-2c3074de3eb9 service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] No waiting events found dispatching network-vif-plugged-a668d6b6-4111-4a94-8bc8-09535b29f895 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1097.757938] env[62692]: WARNING nova.compute.manager [req-4fafd5ce-576e-481f-b9f6-640da3cf5bf1 req-791b0683-6c21-4ab6-972e-2c3074de3eb9 service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Received unexpected event network-vif-plugged-a668d6b6-4111-4a94-8bc8-09535b29f895 for instance with vm_state building and task_state spawning. [ 1097.810302] env[62692]: DEBUG oslo_vmware.api [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141820, 'name': ReconfigVM_Task, 'duration_secs': 0.653324} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.810596] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 1d9271aa-d41d-48cd-a879-3306e7212197/1d9271aa-d41d-48cd-a879-3306e7212197.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1097.811269] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8f417fca-47cc-46b5-84a4-637286e4aa9d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.816954] env[62692]: DEBUG oslo_vmware.api [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1097.816954] env[62692]: value = "task-1141823" [ 1097.816954] env[62692]: _type = "Task" [ 1097.816954] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.821591] env[62692]: DEBUG oslo_concurrency.lockutils [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.313s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.822092] env[62692]: DEBUG nova.compute.manager [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1097.827360] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.574s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.827584] env[62692]: DEBUG nova.objects.instance [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lazy-loading 'resources' on Instance uuid 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1097.829564] env[62692]: DEBUG oslo_vmware.api [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141823, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.836231] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cb7ff3cb-3a9d-4d8e-907e-4676d6ecc8ae tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "refresh_cache-7183ec4a-992d-48f1-8dda-7f499c2f4e1d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1097.836453] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cb7ff3cb-3a9d-4d8e-907e-4676d6ecc8ae tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquired lock "refresh_cache-7183ec4a-992d-48f1-8dda-7f499c2f4e1d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.836564] env[62692]: DEBUG nova.network.neutron [None req-cb7ff3cb-3a9d-4d8e-907e-4676d6ecc8ae tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1097.836740] env[62692]: DEBUG nova.objects.instance [None req-cb7ff3cb-3a9d-4d8e-907e-4676d6ecc8ae tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lazy-loading 'info_cache' on Instance uuid 7183ec4a-992d-48f1-8dda-7f499c2f4e1d {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1097.879321] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141822, 'name': Rename_Task, 'duration_secs': 0.326026} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.879605] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1097.879844] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f8030da9-7cac-4b40-a5c9-25e95b37f2c8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.885755] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1097.885755] env[62692]: value = "task-1141824" [ 1097.885755] env[62692]: _type = "Task" [ 1097.885755] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.893209] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141824, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.990069] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "f92fae48-7285-4b69-97c9-2360227c18fc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.990634] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "f92fae48-7285-4b69-97c9-2360227c18fc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.990884] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "f92fae48-7285-4b69-97c9-2360227c18fc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.991126] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "f92fae48-7285-4b69-97c9-2360227c18fc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.991311] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "f92fae48-7285-4b69-97c9-2360227c18fc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.993506] env[62692]: INFO nova.compute.manager [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Terminating instance [ 1097.995376] env[62692]: DEBUG nova.compute.manager [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1097.995584] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1097.996472] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcc25b43-cb89-4965-adea-4cdfb81b8047 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.004412] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1098.004664] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f85acfe1-934f-47f5-9de4-0e3dd21aaead {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.302648] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1098.302869] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1098.303073] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Deleting the datastore file [datastore2] f92fae48-7285-4b69-97c9-2360227c18fc {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1098.303340] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5797f12b-3da3-4181-bb57-6d605a052b96 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.310166] env[62692]: DEBUG oslo_vmware.api [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for the task: (returnval){ [ 1098.310166] env[62692]: value = "task-1141826" [ 1098.310166] env[62692]: _type = "Task" [ 1098.310166] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.318400] env[62692]: DEBUG oslo_vmware.api [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141826, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.319305] env[62692]: DEBUG nova.network.neutron [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Successfully updated port: a668d6b6-4111-4a94-8bc8-09535b29f895 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1098.333979] env[62692]: DEBUG nova.compute.utils [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1098.335448] env[62692]: DEBUG oslo_vmware.api [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141823, 'name': Rename_Task, 'duration_secs': 0.141947} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.337350] env[62692]: DEBUG nova.compute.manager [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1098.337350] env[62692]: DEBUG nova.network.neutron [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1098.339072] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1098.340887] env[62692]: DEBUG nova.objects.base [None req-cb7ff3cb-3a9d-4d8e-907e-4676d6ecc8ae tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Object Instance<7183ec4a-992d-48f1-8dda-7f499c2f4e1d> lazy-loaded attributes: flavor,info_cache {{(pid=62692) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1098.342038] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-12892ffb-17df-48ea-b1bb-4544cc383fe7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.347595] env[62692]: DEBUG nova.compute.manager [req-32bb9fe5-11e0-46a0-bab2-460d93edb694 req-c69062db-0abf-425e-bae1-8ad42cc13a07 service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Received event network-changed-a668d6b6-4111-4a94-8bc8-09535b29f895 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1098.347794] env[62692]: DEBUG nova.compute.manager [req-32bb9fe5-11e0-46a0-bab2-460d93edb694 req-c69062db-0abf-425e-bae1-8ad42cc13a07 service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Refreshing instance network info cache due to event network-changed-a668d6b6-4111-4a94-8bc8-09535b29f895. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1098.348068] env[62692]: DEBUG oslo_concurrency.lockutils [req-32bb9fe5-11e0-46a0-bab2-460d93edb694 req-c69062db-0abf-425e-bae1-8ad42cc13a07 service nova] Acquiring lock "refresh_cache-1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1098.348287] env[62692]: DEBUG oslo_concurrency.lockutils [req-32bb9fe5-11e0-46a0-bab2-460d93edb694 req-c69062db-0abf-425e-bae1-8ad42cc13a07 service nova] Acquired lock "refresh_cache-1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1098.348496] env[62692]: DEBUG nova.network.neutron [req-32bb9fe5-11e0-46a0-bab2-460d93edb694 req-c69062db-0abf-425e-bae1-8ad42cc13a07 service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Refreshing network info cache for port a668d6b6-4111-4a94-8bc8-09535b29f895 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1098.352385] env[62692]: DEBUG oslo_vmware.api [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1098.352385] env[62692]: value = "task-1141827" [ 1098.352385] env[62692]: _type = "Task" [ 1098.352385] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.363944] env[62692]: DEBUG oslo_vmware.api [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141827, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.385967] env[62692]: DEBUG nova.policy [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b40c8976b544c059458aa4e0dbab160', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9cb8ae08824245a7911de7455e7fcb68', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1098.398368] env[62692]: DEBUG oslo_vmware.api [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141824, 'name': PowerOnVM_Task, 'duration_secs': 0.478553} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.398368] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1098.398368] env[62692]: DEBUG nova.compute.manager [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1098.399279] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c82d8b98-78e6-4226-a169-3a73106e8f5f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.532393] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b69de61a-dce7-4e9f-9532-19bd6f319c99 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.540094] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-265f1de7-2645-4b46-8cb4-1aa5ab1db730 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.571145] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-290f7a0b-1295-4493-ac43-dee19fbda52b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.578983] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2585c1b7-0bec-4c3a-b18c-159846d2063f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.591955] env[62692]: DEBUG nova.compute.provider_tree [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1098.687279] env[62692]: DEBUG nova.network.neutron [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Successfully created port: da5f5c81-a3f5-46d7-b2eb-364476be8845 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1098.822084] env[62692]: DEBUG oslo_vmware.api [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Task: {'id': task-1141826, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165093} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.822352] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1098.822646] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1098.822748] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1098.822947] env[62692]: INFO nova.compute.manager [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Took 0.83 seconds to destroy the instance on the hypervisor. [ 1098.823365] env[62692]: DEBUG oslo.service.loopingcall [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1098.823566] env[62692]: DEBUG nova.compute.manager [-] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1098.823676] env[62692]: DEBUG nova.network.neutron [-] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1098.825903] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "refresh_cache-1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1098.836996] env[62692]: DEBUG nova.compute.manager [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1098.865900] env[62692]: DEBUG oslo_vmware.api [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141827, 'name': PowerOnVM_Task, 'duration_secs': 0.508401} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.867480] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1098.867948] env[62692]: INFO nova.compute.manager [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Took 8.78 seconds to spawn the instance on the hypervisor. [ 1098.867948] env[62692]: DEBUG nova.compute.manager [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1098.868885] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c13ace0-4745-4bb1-b626-1c0c54e1f0b1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.907486] env[62692]: DEBUG nova.network.neutron [req-32bb9fe5-11e0-46a0-bab2-460d93edb694 req-c69062db-0abf-425e-bae1-8ad42cc13a07 service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1098.915405] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.039126] env[62692]: DEBUG nova.network.neutron [req-32bb9fe5-11e0-46a0-bab2-460d93edb694 req-c69062db-0abf-425e-bae1-8ad42cc13a07 service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.095173] env[62692]: DEBUG nova.scheduler.client.report [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1099.276338] env[62692]: DEBUG nova.compute.manager [req-59da89be-cbf4-46fd-84dd-cce78ab54a12 req-655fb2d7-1238-4429-8fe2-de9a63e800c9 service nova] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Received event network-vif-deleted-3c66a7e2-2738-4873-b748-348237876eda {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1099.276526] env[62692]: INFO nova.compute.manager [req-59da89be-cbf4-46fd-84dd-cce78ab54a12 req-655fb2d7-1238-4429-8fe2-de9a63e800c9 service nova] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Neutron deleted interface 3c66a7e2-2738-4873-b748-348237876eda; detaching it from the instance and deleting it from the info cache [ 1099.276696] env[62692]: DEBUG nova.network.neutron [req-59da89be-cbf4-46fd-84dd-cce78ab54a12 req-655fb2d7-1238-4429-8fe2-de9a63e800c9 service nova] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.295177] env[62692]: DEBUG nova.network.neutron [None req-cb7ff3cb-3a9d-4d8e-907e-4676d6ecc8ae tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Updating instance_info_cache with network_info: [{"id": "ab9d5209-bc8f-437d-a219-7f521cae10e6", "address": "fa:16:3e:d1:60:11", "network": {"id": "26fe2f12-b216-4fa5-a022-24c5ab315416", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-838584488-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb2ee16ec90c48e6babfb1d31ea3de94", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab9d5209-bc", "ovs_interfaceid": "ab9d5209-bc8f-437d-a219-7f521cae10e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.385173] env[62692]: INFO nova.compute.manager [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Took 16.07 seconds to build instance. [ 1099.544886] env[62692]: DEBUG oslo_concurrency.lockutils [req-32bb9fe5-11e0-46a0-bab2-460d93edb694 req-c69062db-0abf-425e-bae1-8ad42cc13a07 service nova] Releasing lock "refresh_cache-1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1099.545325] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "refresh_cache-1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1099.545493] env[62692]: DEBUG nova.network.neutron [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1099.599743] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.772s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.602142] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.536s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.602395] env[62692]: DEBUG nova.objects.instance [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Lazy-loading 'resources' on Instance uuid f27e3a4f-d109-472b-a348-5099bdc0e7c8 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1099.626553] env[62692]: INFO nova.scheduler.client.report [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Deleted allocations for instance 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da [ 1099.750079] env[62692]: DEBUG nova.network.neutron [-] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.779672] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4573780f-6275-4762-ba1a-b1f43f3eb994 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.790614] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58865098-fc54-49df-92d2-6a3e799674ea {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.803941] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cb7ff3cb-3a9d-4d8e-907e-4676d6ecc8ae tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Releasing lock "refresh_cache-7183ec4a-992d-48f1-8dda-7f499c2f4e1d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1099.823822] env[62692]: DEBUG nova.compute.manager [req-59da89be-cbf4-46fd-84dd-cce78ab54a12 req-655fb2d7-1238-4429-8fe2-de9a63e800c9 service nova] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Detach interface failed, port_id=3c66a7e2-2738-4873-b748-348237876eda, reason: Instance f92fae48-7285-4b69-97c9-2360227c18fc could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1099.847216] env[62692]: DEBUG nova.compute.manager [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1099.872738] env[62692]: DEBUG nova.virt.hardware [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1099.872738] env[62692]: DEBUG nova.virt.hardware [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1099.872738] env[62692]: DEBUG nova.virt.hardware [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1099.873469] env[62692]: DEBUG nova.virt.hardware [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1099.873469] env[62692]: DEBUG nova.virt.hardware [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1099.873469] env[62692]: DEBUG nova.virt.hardware [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1099.873469] env[62692]: DEBUG nova.virt.hardware [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1099.873630] env[62692]: DEBUG nova.virt.hardware [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1099.873659] env[62692]: DEBUG nova.virt.hardware [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1099.873870] env[62692]: DEBUG nova.virt.hardware [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1099.874075] env[62692]: DEBUG nova.virt.hardware [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1099.875053] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d70245-ad41-487b-929b-bc3f692a04fb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.883679] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ed7e6cf-79de-4d81-8356-89adb088fc65 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.888906] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab39cda8-dc54-4d97-bd3c-15a30f2a1159 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "1d9271aa-d41d-48cd-a879-3306e7212197" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.580s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.080482] env[62692]: DEBUG nova.network.neutron [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1100.134111] env[62692]: DEBUG oslo_concurrency.lockutils [None req-96e0e3f7-3343-433a-81f3-3c7388b224ef tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.362s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.224018] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4566186-31ec-4b64-a557-a80bf89509d1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.227933] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-55f6e542-c4ba-4724-945c-e95bab23c95c tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Suspending the VM {{(pid=62692) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1100.228468] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-861fa19c-0929-4743-bb8c-773dd8d5dd1f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.234794] env[62692]: DEBUG oslo_vmware.api [None req-55f6e542-c4ba-4724-945c-e95bab23c95c tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1100.234794] env[62692]: value = "task-1141828" [ 1100.234794] env[62692]: _type = "Task" [ 1100.234794] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.241973] env[62692]: DEBUG nova.network.neutron [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Updating instance_info_cache with network_info: [{"id": "a668d6b6-4111-4a94-8bc8-09535b29f895", "address": "fa:16:3e:02:7f:2a", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa668d6b6-41", "ovs_interfaceid": "a668d6b6-4111-4a94-8bc8-09535b29f895", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1100.250090] env[62692]: DEBUG oslo_vmware.api [None req-55f6e542-c4ba-4724-945c-e95bab23c95c tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141828, 'name': SuspendVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.252626] env[62692]: INFO nova.compute.manager [-] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Took 1.43 seconds to deallocate network for instance. [ 1100.299345] env[62692]: DEBUG nova.network.neutron [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Successfully updated port: da5f5c81-a3f5-46d7-b2eb-364476be8845 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1100.305982] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1683fb0d-d1eb-43dc-a8b0-201fc731b5a4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.309691] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb7ff3cb-3a9d-4d8e-907e-4676d6ecc8ae tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1100.309992] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-71a7e42f-6eb0-4650-ab30-ad0e8fb699ba {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.319515] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e171fe41-cc1f-4516-bd33-460b11cd487e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.323209] env[62692]: DEBUG oslo_vmware.api [None req-cb7ff3cb-3a9d-4d8e-907e-4676d6ecc8ae tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1100.323209] env[62692]: value = "task-1141829" [ 1100.323209] env[62692]: _type = "Task" [ 1100.323209] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.357712] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3144f59-da92-4891-929a-ba1293986267 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.360570] env[62692]: DEBUG oslo_vmware.api [None req-cb7ff3cb-3a9d-4d8e-907e-4676d6ecc8ae tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141829, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.365769] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7738606-0a43-4d83-91fe-c3ed82e01659 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.380927] env[62692]: DEBUG nova.compute.provider_tree [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1100.383786] env[62692]: DEBUG nova.compute.manager [req-15437542-9e13-49a1-96d3-37f69324b760 req-fceb3387-b67d-48ef-bf04-eaeebfb8c461 service nova] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Received event network-vif-plugged-da5f5c81-a3f5-46d7-b2eb-364476be8845 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1100.384017] env[62692]: DEBUG oslo_concurrency.lockutils [req-15437542-9e13-49a1-96d3-37f69324b760 req-fceb3387-b67d-48ef-bf04-eaeebfb8c461 service nova] Acquiring lock "adfb0194-e94b-4a87-8554-a5c262fe9a70-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.384231] env[62692]: DEBUG oslo_concurrency.lockutils [req-15437542-9e13-49a1-96d3-37f69324b760 req-fceb3387-b67d-48ef-bf04-eaeebfb8c461 service nova] Lock "adfb0194-e94b-4a87-8554-a5c262fe9a70-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.384405] env[62692]: DEBUG oslo_concurrency.lockutils [req-15437542-9e13-49a1-96d3-37f69324b760 req-fceb3387-b67d-48ef-bf04-eaeebfb8c461 service nova] Lock "adfb0194-e94b-4a87-8554-a5c262fe9a70-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.384574] env[62692]: DEBUG nova.compute.manager [req-15437542-9e13-49a1-96d3-37f69324b760 req-fceb3387-b67d-48ef-bf04-eaeebfb8c461 service nova] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] No waiting events found dispatching network-vif-plugged-da5f5c81-a3f5-46d7-b2eb-364476be8845 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1100.384741] env[62692]: WARNING nova.compute.manager [req-15437542-9e13-49a1-96d3-37f69324b760 req-fceb3387-b67d-48ef-bf04-eaeebfb8c461 service nova] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Received unexpected event network-vif-plugged-da5f5c81-a3f5-46d7-b2eb-364476be8845 for instance with vm_state building and task_state spawning. [ 1100.384901] env[62692]: DEBUG nova.compute.manager [req-15437542-9e13-49a1-96d3-37f69324b760 req-fceb3387-b67d-48ef-bf04-eaeebfb8c461 service nova] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Received event network-changed-da5f5c81-a3f5-46d7-b2eb-364476be8845 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1100.385074] env[62692]: DEBUG nova.compute.manager [req-15437542-9e13-49a1-96d3-37f69324b760 req-fceb3387-b67d-48ef-bf04-eaeebfb8c461 service nova] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Refreshing instance network info cache due to event network-changed-da5f5c81-a3f5-46d7-b2eb-364476be8845. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1100.385264] env[62692]: DEBUG oslo_concurrency.lockutils [req-15437542-9e13-49a1-96d3-37f69324b760 req-fceb3387-b67d-48ef-bf04-eaeebfb8c461 service nova] Acquiring lock "refresh_cache-adfb0194-e94b-4a87-8554-a5c262fe9a70" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1100.385531] env[62692]: DEBUG oslo_concurrency.lockutils [req-15437542-9e13-49a1-96d3-37f69324b760 req-fceb3387-b67d-48ef-bf04-eaeebfb8c461 service nova] Acquired lock "refresh_cache-adfb0194-e94b-4a87-8554-a5c262fe9a70" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.385585] env[62692]: DEBUG nova.network.neutron [req-15437542-9e13-49a1-96d3-37f69324b760 req-fceb3387-b67d-48ef-bf04-eaeebfb8c461 service nova] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Refreshing network info cache for port da5f5c81-a3f5-46d7-b2eb-364476be8845 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1100.584753] env[62692]: DEBUG oslo_concurrency.lockutils [None req-63db180d-1e26-4307-9b11-a433acb9d4fa tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "e8f645da-3b4a-4f21-9a17-aebdc8d112df" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.585278] env[62692]: DEBUG oslo_concurrency.lockutils [None req-63db180d-1e26-4307-9b11-a433acb9d4fa tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "e8f645da-3b4a-4f21-9a17-aebdc8d112df" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.744717] env[62692]: DEBUG oslo_vmware.api [None req-55f6e542-c4ba-4724-945c-e95bab23c95c tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141828, 'name': SuspendVM_Task} progress is 62%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.746270] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "refresh_cache-1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1100.746565] env[62692]: DEBUG nova.compute.manager [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Instance network_info: |[{"id": "a668d6b6-4111-4a94-8bc8-09535b29f895", "address": "fa:16:3e:02:7f:2a", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa668d6b6-41", "ovs_interfaceid": "a668d6b6-4111-4a94-8bc8-09535b29f895", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1100.746968] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:02:7f:2a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6a31b224-77d7-48a4-af87-312758611b8e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a668d6b6-4111-4a94-8bc8-09535b29f895', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1100.754556] env[62692]: DEBUG oslo.service.loopingcall [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1100.754791] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1100.755031] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4008c49e-d17e-4c9c-aa23-82056194a3e2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.770461] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.775985] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1100.775985] env[62692]: value = "task-1141830" [ 1100.775985] env[62692]: _type = "Task" [ 1100.775985] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.785300] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141830, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.801171] env[62692]: DEBUG oslo_concurrency.lockutils [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "refresh_cache-adfb0194-e94b-4a87-8554-a5c262fe9a70" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1100.833969] env[62692]: DEBUG oslo_vmware.api [None req-cb7ff3cb-3a9d-4d8e-907e-4676d6ecc8ae tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141829, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.887863] env[62692]: DEBUG nova.scheduler.client.report [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1100.924850] env[62692]: DEBUG nova.network.neutron [req-15437542-9e13-49a1-96d3-37f69324b760 req-fceb3387-b67d-48ef-bf04-eaeebfb8c461 service nova] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1101.010822] env[62692]: DEBUG nova.network.neutron [req-15437542-9e13-49a1-96d3-37f69324b760 req-fceb3387-b67d-48ef-bf04-eaeebfb8c461 service nova] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.089223] env[62692]: INFO nova.compute.manager [None req-63db180d-1e26-4307-9b11-a433acb9d4fa tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Detaching volume ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60 [ 1101.131785] env[62692]: INFO nova.virt.block_device [None req-63db180d-1e26-4307-9b11-a433acb9d4fa tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Attempting to driver detach volume ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60 from mountpoint /dev/sdb [ 1101.132063] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-63db180d-1e26-4307-9b11-a433acb9d4fa tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Volume detach. Driver type: vmdk {{(pid=62692) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1101.132279] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-63db180d-1e26-4307-9b11-a433acb9d4fa tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249042', 'volume_id': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'name': 'volume-ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e8f645da-3b4a-4f21-9a17-aebdc8d112df', 'attached_at': '', 'detached_at': '', 'volume_id': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'serial': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1101.133279] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0444b2a0-d359-4719-863b-c262c161af06 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.158150] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-294b682d-0f04-498b-8c17-2a95b7bf9f2a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.165561] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64b7f625-c17d-4355-bb38-aec5faf1f880 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.184993] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebf824b4-2271-4beb-9192-0f56365e470a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.199221] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-63db180d-1e26-4307-9b11-a433acb9d4fa tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] The volume has not been displaced from its original location: [datastore2] volume-ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60/volume-ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60.vmdk. No consolidation needed. {{(pid=62692) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1101.204305] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-63db180d-1e26-4307-9b11-a433acb9d4fa tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Reconfiguring VM instance instance-00000059 to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1101.204601] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-702ffeea-74c6-4e36-b8bf-35896334c156 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.222076] env[62692]: DEBUG oslo_vmware.api [None req-63db180d-1e26-4307-9b11-a433acb9d4fa tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1101.222076] env[62692]: value = "task-1141831" [ 1101.222076] env[62692]: _type = "Task" [ 1101.222076] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.230401] env[62692]: DEBUG oslo_vmware.api [None req-63db180d-1e26-4307-9b11-a433acb9d4fa tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141831, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.243336] env[62692]: DEBUG oslo_vmware.api [None req-55f6e542-c4ba-4724-945c-e95bab23c95c tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141828, 'name': SuspendVM_Task, 'duration_secs': 0.736036} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.243676] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-55f6e542-c4ba-4724-945c-e95bab23c95c tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Suspended the VM {{(pid=62692) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1101.243900] env[62692]: DEBUG nova.compute.manager [None req-55f6e542-c4ba-4724-945c-e95bab23c95c tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1101.244729] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d8f6e0-fd4e-432e-aab9-bd842c400c93 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.285173] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141830, 'name': CreateVM_Task, 'duration_secs': 0.398834} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.285969] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1101.286682] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1101.286851] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.287186] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1101.287672] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1bd38d0-e6cf-4b5d-9ca3-83784b6e607f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.292152] env[62692]: DEBUG oslo_vmware.api [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1101.292152] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52915a7b-32f7-c3e5-14a4-e5a2fcde823a" [ 1101.292152] env[62692]: _type = "Task" [ 1101.292152] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.298881] env[62692]: DEBUG oslo_vmware.api [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52915a7b-32f7-c3e5-14a4-e5a2fcde823a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.334182] env[62692]: DEBUG oslo_vmware.api [None req-cb7ff3cb-3a9d-4d8e-907e-4676d6ecc8ae tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141829, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.395758] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.794s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.398169] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 2.483s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.398377] env[62692]: DEBUG nova.objects.instance [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62692) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1101.418382] env[62692]: INFO nova.scheduler.client.report [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Deleted allocations for instance f27e3a4f-d109-472b-a348-5099bdc0e7c8 [ 1101.514083] env[62692]: DEBUG oslo_concurrency.lockutils [req-15437542-9e13-49a1-96d3-37f69324b760 req-fceb3387-b67d-48ef-bf04-eaeebfb8c461 service nova] Releasing lock "refresh_cache-adfb0194-e94b-4a87-8554-a5c262fe9a70" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1101.514495] env[62692]: DEBUG oslo_concurrency.lockutils [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired lock "refresh_cache-adfb0194-e94b-4a87-8554-a5c262fe9a70" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.514659] env[62692]: DEBUG nova.network.neutron [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1101.731688] env[62692]: DEBUG oslo_vmware.api [None req-63db180d-1e26-4307-9b11-a433acb9d4fa tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141831, 'name': ReconfigVM_Task, 'duration_secs': 0.334166} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.731985] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-63db180d-1e26-4307-9b11-a433acb9d4fa tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Reconfigured VM instance instance-00000059 to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1101.736685] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1a58b7c0-6909-43d9-b732-190afbf7509b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.751692] env[62692]: DEBUG oslo_vmware.api [None req-63db180d-1e26-4307-9b11-a433acb9d4fa tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1101.751692] env[62692]: value = "task-1141832" [ 1101.751692] env[62692]: _type = "Task" [ 1101.751692] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.762384] env[62692]: DEBUG oslo_vmware.api [None req-63db180d-1e26-4307-9b11-a433acb9d4fa tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141832, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.802950] env[62692]: DEBUG oslo_vmware.api [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52915a7b-32f7-c3e5-14a4-e5a2fcde823a, 'name': SearchDatastore_Task, 'duration_secs': 0.018799} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.803308] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1101.803925] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1101.803925] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1101.803925] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.804114] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1101.804703] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c8415983-38ca-4e49-8515-e478c1693ebc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.814659] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1101.814834] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1101.815594] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27668c9e-0397-449b-b329-eb94f98a41b6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.820540] env[62692]: DEBUG oslo_vmware.api [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1101.820540] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]523de1f8-de73-4f04-954c-89279093e62a" [ 1101.820540] env[62692]: _type = "Task" [ 1101.820540] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.827907] env[62692]: DEBUG oslo_vmware.api [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]523de1f8-de73-4f04-954c-89279093e62a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.834734] env[62692]: DEBUG oslo_vmware.api [None req-cb7ff3cb-3a9d-4d8e-907e-4676d6ecc8ae tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141829, 'name': PowerOnVM_Task, 'duration_secs': 1.040043} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.835015] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb7ff3cb-3a9d-4d8e-907e-4676d6ecc8ae tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1101.835228] env[62692]: DEBUG nova.compute.manager [None req-cb7ff3cb-3a9d-4d8e-907e-4676d6ecc8ae tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1101.835921] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-879e3ee4-86c7-490c-9b3a-77be66629546 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.926214] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8bb5e2cc-27ea-4d43-aa71-7711de789141 tempest-ServerMetadataTestJSON-87048793 tempest-ServerMetadataTestJSON-87048793-project-member] Lock "f27e3a4f-d109-472b-a348-5099bdc0e7c8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.076s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.057354] env[62692]: DEBUG nova.network.neutron [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1102.227264] env[62692]: DEBUG nova.network.neutron [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Updating instance_info_cache with network_info: [{"id": "da5f5c81-a3f5-46d7-b2eb-364476be8845", "address": "fa:16:3e:81:a0:c0", "network": {"id": "f6352999-f943-431d-b54a-c9dc1d810a85", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-121037111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9cb8ae08824245a7911de7455e7fcb68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c842425c-544e-4ce2-9657-512723bd318e", "external-id": "nsx-vlan-transportzone-80", "segmentation_id": 80, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda5f5c81-a3", "ovs_interfaceid": "da5f5c81-a3f5-46d7-b2eb-364476be8845", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1102.261885] env[62692]: DEBUG oslo_vmware.api [None req-63db180d-1e26-4307-9b11-a433acb9d4fa tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141832, 'name': ReconfigVM_Task, 'duration_secs': 0.190204} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.262214] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-63db180d-1e26-4307-9b11-a433acb9d4fa tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249042', 'volume_id': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'name': 'volume-ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e8f645da-3b4a-4f21-9a17-aebdc8d112df', 'attached_at': '', 'detached_at': '', 'volume_id': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60', 'serial': 'ccbac0fe-58e0-4a1e-8ed3-74fcc90c8e60'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1102.330962] env[62692]: DEBUG oslo_vmware.api [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]523de1f8-de73-4f04-954c-89279093e62a, 'name': SearchDatastore_Task, 'duration_secs': 0.034451} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.331782] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c0f2c803-2057-45d6-91a4-b13efd528bdc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.336743] env[62692]: DEBUG oslo_vmware.api [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1102.336743] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52c2386e-4708-b7af-2210-2888ab5dc290" [ 1102.336743] env[62692]: _type = "Task" [ 1102.336743] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.347172] env[62692]: DEBUG oslo_vmware.api [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52c2386e-4708-b7af-2210-2888ab5dc290, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.408433] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cd23d59e-3c1f-45a4-889c-6453eecc42d8 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.409642] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.639s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.409888] env[62692]: DEBUG nova.objects.instance [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lazy-loading 'resources' on Instance uuid f92fae48-7285-4b69-97c9-2360227c18fc {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1102.732620] env[62692]: DEBUG oslo_concurrency.lockutils [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Releasing lock "refresh_cache-adfb0194-e94b-4a87-8554-a5c262fe9a70" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1102.732975] env[62692]: DEBUG nova.compute.manager [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Instance network_info: |[{"id": "da5f5c81-a3f5-46d7-b2eb-364476be8845", "address": "fa:16:3e:81:a0:c0", "network": {"id": "f6352999-f943-431d-b54a-c9dc1d810a85", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-121037111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9cb8ae08824245a7911de7455e7fcb68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c842425c-544e-4ce2-9657-512723bd318e", "external-id": "nsx-vlan-transportzone-80", "segmentation_id": 80, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda5f5c81-a3", "ovs_interfaceid": "da5f5c81-a3f5-46d7-b2eb-364476be8845", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1102.733426] env[62692]: DEBUG oslo_concurrency.lockutils [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "1d9271aa-d41d-48cd-a879-3306e7212197" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.733667] env[62692]: DEBUG oslo_concurrency.lockutils [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "1d9271aa-d41d-48cd-a879-3306e7212197" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.734396] env[62692]: DEBUG oslo_concurrency.lockutils [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "1d9271aa-d41d-48cd-a879-3306e7212197-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.734396] env[62692]: DEBUG oslo_concurrency.lockutils [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "1d9271aa-d41d-48cd-a879-3306e7212197-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.734396] env[62692]: DEBUG oslo_concurrency.lockutils [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "1d9271aa-d41d-48cd-a879-3306e7212197-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.736025] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:a0:c0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c842425c-544e-4ce2-9657-512723bd318e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'da5f5c81-a3f5-46d7-b2eb-364476be8845', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1102.743660] env[62692]: DEBUG oslo.service.loopingcall [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1102.744172] env[62692]: INFO nova.compute.manager [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Terminating instance [ 1102.745727] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1102.745979] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-620ec5e0-c77d-45bd-8add-3fbe7d379215 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.763029] env[62692]: DEBUG nova.compute.manager [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1102.763289] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1102.764560] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6bac50c-9032-401b-aa2b-673dffbbb916 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.773376] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1102.774882] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e030c802-7369-4f60-8be4-67203f8e9d43 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.776215] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1102.776215] env[62692]: value = "task-1141833" [ 1102.776215] env[62692]: _type = "Task" [ 1102.776215] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.784493] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141833, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.807018] env[62692]: DEBUG nova.objects.instance [None req-63db180d-1e26-4307-9b11-a433acb9d4fa tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lazy-loading 'flavor' on Instance uuid e8f645da-3b4a-4f21-9a17-aebdc8d112df {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1102.843778] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1102.844194] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1102.844525] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Deleting the datastore file [datastore1] 1d9271aa-d41d-48cd-a879-3306e7212197 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1102.844925] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1a50d423-e227-4bed-ba41-7efbcb217cc8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.856829] env[62692]: DEBUG oslo_vmware.api [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52c2386e-4708-b7af-2210-2888ab5dc290, 'name': SearchDatastore_Task, 'duration_secs': 0.01428} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.858833] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1102.859282] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 1904edb2-fd72-43b6-9219-8043c5abff0c/1904edb2-fd72-43b6-9219-8043c5abff0c.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1102.859794] env[62692]: DEBUG oslo_vmware.api [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1102.859794] env[62692]: value = "task-1141835" [ 1102.859794] env[62692]: _type = "Task" [ 1102.859794] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.860114] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1f43374f-6186-4cf9-b209-196e4ea2e9d2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.874608] env[62692]: DEBUG oslo_vmware.api [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141835, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.876377] env[62692]: DEBUG oslo_vmware.api [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1102.876377] env[62692]: value = "task-1141836" [ 1102.876377] env[62692]: _type = "Task" [ 1102.876377] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.886941] env[62692]: DEBUG oslo_vmware.api [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141836, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.088977] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66bd3708-01df-4602-ae2b-17680e479cd7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.098039] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-254f0376-dcc6-40f7-816b-000fe9897641 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.134020] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9182ff09-8558-4803-b361-1fe3cde5d26d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.142144] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bba11e56-4fa6-465d-aa68-feb065b3ab62 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.158533] env[62692]: DEBUG nova.compute.provider_tree [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1103.290532] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141833, 'name': CreateVM_Task, 'duration_secs': 0.312931} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.290654] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1103.291396] env[62692]: DEBUG oslo_concurrency.lockutils [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1103.291563] env[62692]: DEBUG oslo_concurrency.lockutils [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.291910] env[62692]: DEBUG oslo_concurrency.lockutils [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1103.292226] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54753251-9b9d-40a6-b2fe-d77c7212b3a8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.298340] env[62692]: DEBUG oslo_vmware.api [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1103.298340] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]525670a9-4f46-b39f-64b7-0edbdb6ae593" [ 1103.298340] env[62692]: _type = "Task" [ 1103.298340] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.308357] env[62692]: DEBUG oslo_vmware.api [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]525670a9-4f46-b39f-64b7-0edbdb6ae593, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.376292] env[62692]: DEBUG oslo_vmware.api [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141835, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169756} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.376292] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1103.376292] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1103.376292] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1103.376292] env[62692]: INFO nova.compute.manager [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Took 0.61 seconds to destroy the instance on the hypervisor. [ 1103.376292] env[62692]: DEBUG oslo.service.loopingcall [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1103.376292] env[62692]: DEBUG nova.compute.manager [-] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1103.376292] env[62692]: DEBUG nova.network.neutron [-] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1103.386116] env[62692]: DEBUG oslo_vmware.api [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141836, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.662622] env[62692]: DEBUG nova.scheduler.client.report [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1103.813919] env[62692]: DEBUG oslo_vmware.api [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]525670a9-4f46-b39f-64b7-0edbdb6ae593, 'name': SearchDatastore_Task, 'duration_secs': 0.020488} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.813919] env[62692]: DEBUG oslo_concurrency.lockutils [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1103.813919] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1103.813919] env[62692]: DEBUG oslo_concurrency.lockutils [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1103.813919] env[62692]: DEBUG oslo_concurrency.lockutils [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.813919] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1103.813919] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3b98a209-8271-4af7-8c7e-a683319effa5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.817059] env[62692]: DEBUG oslo_concurrency.lockutils [None req-63db180d-1e26-4307-9b11-a433acb9d4fa tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "e8f645da-3b4a-4f21-9a17-aebdc8d112df" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.232s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.828027] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1103.828027] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1103.828027] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02050227-99e4-4361-9ba1-0a71382f5a51 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.832360] env[62692]: DEBUG oslo_vmware.api [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1103.832360] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]528b5210-c8e1-7e31-05a4-7406e0f1a58e" [ 1103.832360] env[62692]: _type = "Task" [ 1103.832360] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.841412] env[62692]: DEBUG oslo_vmware.api [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]528b5210-c8e1-7e31-05a4-7406e0f1a58e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.887732] env[62692]: DEBUG oslo_vmware.api [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141836, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.517685} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.888425] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 1904edb2-fd72-43b6-9219-8043c5abff0c/1904edb2-fd72-43b6-9219-8043c5abff0c.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1103.888698] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1103.888962] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-50928548-2b07-4304-b48c-7b60203dc250 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.896862] env[62692]: DEBUG oslo_vmware.api [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1103.896862] env[62692]: value = "task-1141837" [ 1103.896862] env[62692]: _type = "Task" [ 1103.896862] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.901572] env[62692]: DEBUG nova.compute.manager [req-07615a8a-fb34-4552-900f-696d2dc37537 req-f52f6852-0331-4086-80e7-25e6691d8763 service nova] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Received event network-vif-deleted-65f2688b-0a0c-486b-a629-e9dc96ad98b6 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1103.901775] env[62692]: INFO nova.compute.manager [req-07615a8a-fb34-4552-900f-696d2dc37537 req-f52f6852-0331-4086-80e7-25e6691d8763 service nova] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Neutron deleted interface 65f2688b-0a0c-486b-a629-e9dc96ad98b6; detaching it from the instance and deleting it from the info cache [ 1103.901956] env[62692]: DEBUG nova.network.neutron [req-07615a8a-fb34-4552-900f-696d2dc37537 req-f52f6852-0331-4086-80e7-25e6691d8763 service nova] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1103.910935] env[62692]: DEBUG oslo_vmware.api [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141837, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.167856] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.758s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.197353] env[62692]: INFO nova.scheduler.client.report [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Deleted allocations for instance f92fae48-7285-4b69-97c9-2360227c18fc [ 1104.244826] env[62692]: DEBUG nova.network.neutron [-] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.349150] env[62692]: DEBUG oslo_vmware.api [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]528b5210-c8e1-7e31-05a4-7406e0f1a58e, 'name': SearchDatastore_Task, 'duration_secs': 0.019518} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.350044] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-36da02df-ea20-4bee-bfa1-bffc197e10f9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.361420] env[62692]: DEBUG oslo_vmware.api [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1104.361420] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]525e54ce-3932-3c6d-ba1f-9853c775839a" [ 1104.361420] env[62692]: _type = "Task" [ 1104.361420] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.380620] env[62692]: DEBUG oslo_vmware.api [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]525e54ce-3932-3c6d-ba1f-9853c775839a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.407187] env[62692]: DEBUG oslo_vmware.api [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141837, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.152939} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.407447] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-90579b47-2ef6-4571-ab02-2ff6c82d1365 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.409499] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1104.411061] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a229cba8-cd82-4617-ad3a-eb803efbf9c8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.429117] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e633190f-3b92-49dc-a62a-da86e8dbd83e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.449028] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 1904edb2-fd72-43b6-9219-8043c5abff0c/1904edb2-fd72-43b6-9219-8043c5abff0c.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1104.449729] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e363d7f1-db29-4c6e-a714-ab18b741c110 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.468732] env[62692]: DEBUG oslo_vmware.api [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1104.468732] env[62692]: value = "task-1141838" [ 1104.468732] env[62692]: _type = "Task" [ 1104.468732] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.476500] env[62692]: DEBUG oslo_vmware.api [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141838, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.486515] env[62692]: DEBUG nova.compute.manager [req-07615a8a-fb34-4552-900f-696d2dc37537 req-f52f6852-0331-4086-80e7-25e6691d8763 service nova] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Detach interface failed, port_id=65f2688b-0a0c-486b-a629-e9dc96ad98b6, reason: Instance 1d9271aa-d41d-48cd-a879-3306e7212197 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1104.705520] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6d1a7503-71a7-4d10-a5bf-045fe23bd044 tempest-ServersTestJSON-2071851673 tempest-ServersTestJSON-2071851673-project-member] Lock "f92fae48-7285-4b69-97c9-2360227c18fc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.715s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.748415] env[62692]: INFO nova.compute.manager [-] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Took 1.37 seconds to deallocate network for instance. [ 1104.873601] env[62692]: DEBUG oslo_vmware.api [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]525e54ce-3932-3c6d-ba1f-9853c775839a, 'name': SearchDatastore_Task, 'duration_secs': 0.042934} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.873864] env[62692]: DEBUG oslo_concurrency.lockutils [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.875059] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] adfb0194-e94b-4a87-8554-a5c262fe9a70/adfb0194-e94b-4a87-8554-a5c262fe9a70.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1104.875059] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-642c142a-fb95-42c7-a046-7264ffb3abc1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.876413] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "e8f645da-3b4a-4f21-9a17-aebdc8d112df" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.876639] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "e8f645da-3b4a-4f21-9a17-aebdc8d112df" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.876855] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "e8f645da-3b4a-4f21-9a17-aebdc8d112df-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.878046] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "e8f645da-3b4a-4f21-9a17-aebdc8d112df-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.878046] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "e8f645da-3b4a-4f21-9a17-aebdc8d112df-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.879447] env[62692]: INFO nova.compute.manager [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Terminating instance [ 1104.883647] env[62692]: DEBUG oslo_vmware.api [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1104.883647] env[62692]: value = "task-1141839" [ 1104.883647] env[62692]: _type = "Task" [ 1104.883647] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.888042] env[62692]: DEBUG nova.compute.manager [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1104.888042] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1104.888459] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f34399f2-4466-4ada-afd2-3986b6f30693 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.896075] env[62692]: DEBUG oslo_vmware.api [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141839, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.898565] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1104.899076] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4e0205af-e66b-4256-9644-8fed6a8e4feb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.905640] env[62692]: DEBUG oslo_vmware.api [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1104.905640] env[62692]: value = "task-1141840" [ 1104.905640] env[62692]: _type = "Task" [ 1104.905640] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.913692] env[62692]: DEBUG oslo_vmware.api [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141840, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.978817] env[62692]: DEBUG oslo_vmware.api [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141838, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.255577] env[62692]: DEBUG oslo_concurrency.lockutils [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.255875] env[62692]: DEBUG oslo_concurrency.lockutils [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.256166] env[62692]: DEBUG nova.objects.instance [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lazy-loading 'resources' on Instance uuid 1d9271aa-d41d-48cd-a879-3306e7212197 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1105.394526] env[62692]: DEBUG oslo_vmware.api [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141839, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.415387] env[62692]: DEBUG oslo_vmware.api [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141840, 'name': PowerOffVM_Task, 'duration_secs': 0.206949} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.415675] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1105.415879] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1105.416188] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c76390bf-bbc5-489f-b9a3-92ca3618c1d5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.482229] env[62692]: DEBUG oslo_vmware.api [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141838, 'name': ReconfigVM_Task, 'duration_secs': 0.6693} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.483624] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 1904edb2-fd72-43b6-9219-8043c5abff0c/1904edb2-fd72-43b6-9219-8043c5abff0c.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1105.484853] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1105.488021] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1105.488021] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Deleting the datastore file [datastore1] e8f645da-3b4a-4f21-9a17-aebdc8d112df {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1105.488021] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4c735e8b-2f8e-4d40-8bb9-cf2529b0a5d6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.488718] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-15526c08-aa99-4b0d-9502-44c20fafdd92 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.497295] env[62692]: DEBUG oslo_vmware.api [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1105.497295] env[62692]: value = "task-1141842" [ 1105.497295] env[62692]: _type = "Task" [ 1105.497295] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.499359] env[62692]: DEBUG oslo_vmware.api [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1105.499359] env[62692]: value = "task-1141843" [ 1105.499359] env[62692]: _type = "Task" [ 1105.499359] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.512866] env[62692]: DEBUG oslo_vmware.api [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141842, 'name': Rename_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.516838] env[62692]: DEBUG oslo_vmware.api [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141843, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.894473] env[62692]: DEBUG oslo_vmware.api [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141839, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.564533} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.894789] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] adfb0194-e94b-4a87-8554-a5c262fe9a70/adfb0194-e94b-4a87-8554-a5c262fe9a70.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1105.894966] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1105.895227] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4a941418-8a2b-41f2-beea-42169544a63c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.904073] env[62692]: DEBUG oslo_vmware.api [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1105.904073] env[62692]: value = "task-1141844" [ 1105.904073] env[62692]: _type = "Task" [ 1105.904073] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.912297] env[62692]: DEBUG oslo_vmware.api [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141844, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.926538] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebc43fd5-bcfe-46a2-b91a-57ea8969fcb3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.937657] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-161c834e-70e5-4141-9ffa-f6c9ab349698 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.980604] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f927f6-44e6-405c-a49b-5f6ad38236ad {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.989318] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9378354c-dc06-4e58-b764-d4db1b77c60c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.002964] env[62692]: DEBUG nova.compute.provider_tree [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1106.016974] env[62692]: DEBUG oslo_vmware.api [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141843, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.290864} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.021240] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1106.021240] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1106.021240] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1106.021240] env[62692]: INFO nova.compute.manager [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1106.021240] env[62692]: DEBUG oslo.service.loopingcall [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1106.021240] env[62692]: DEBUG oslo_vmware.api [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141842, 'name': Rename_Task, 'duration_secs': 0.288024} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.021240] env[62692]: DEBUG nova.compute.manager [-] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1106.021240] env[62692]: DEBUG nova.network.neutron [-] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1106.022753] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1106.023582] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ec0ebe77-1ded-490c-8a24-17c047c69610 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.030651] env[62692]: DEBUG oslo_vmware.api [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1106.030651] env[62692]: value = "task-1141845" [ 1106.030651] env[62692]: _type = "Task" [ 1106.030651] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.041017] env[62692]: DEBUG oslo_vmware.api [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141845, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.414932] env[62692]: DEBUG oslo_vmware.api [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141844, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078458} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.415476] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1106.416748] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eefe07c5-dc03-4334-ad32-824a3373e794 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.445586] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] adfb0194-e94b-4a87-8554-a5c262fe9a70/adfb0194-e94b-4a87-8554-a5c262fe9a70.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1106.446927] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-42282927-6072-4475-9259-ea5c399a5285 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.469607] env[62692]: DEBUG oslo_vmware.api [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1106.469607] env[62692]: value = "task-1141846" [ 1106.469607] env[62692]: _type = "Task" [ 1106.469607] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.480750] env[62692]: DEBUG oslo_vmware.api [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141846, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.506028] env[62692]: DEBUG nova.scheduler.client.report [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1106.540726] env[62692]: DEBUG oslo_vmware.api [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141845, 'name': PowerOnVM_Task, 'duration_secs': 0.508616} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.541130] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1106.541444] env[62692]: INFO nova.compute.manager [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Took 9.00 seconds to spawn the instance on the hypervisor. [ 1106.541640] env[62692]: DEBUG nova.compute.manager [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1106.543256] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d0b367a-a492-4ff4-ab96-4684563a9db4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.980500] env[62692]: DEBUG oslo_vmware.api [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141846, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.013128] env[62692]: DEBUG oslo_concurrency.lockutils [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.757s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.040904] env[62692]: INFO nova.scheduler.client.report [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Deleted allocations for instance 1d9271aa-d41d-48cd-a879-3306e7212197 [ 1107.063100] env[62692]: INFO nova.compute.manager [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Took 18.38 seconds to build instance. [ 1107.200633] env[62692]: DEBUG nova.compute.manager [req-a49e6360-af9a-435b-bc3d-e9956c47cf85 req-d9c0e90e-c21a-4f53-8b1e-51e6f4ed158f service nova] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Received event network-vif-deleted-877ab637-ea08-499f-9d34-88ca15e1ed1a {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1107.200777] env[62692]: INFO nova.compute.manager [req-a49e6360-af9a-435b-bc3d-e9956c47cf85 req-d9c0e90e-c21a-4f53-8b1e-51e6f4ed158f service nova] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Neutron deleted interface 877ab637-ea08-499f-9d34-88ca15e1ed1a; detaching it from the instance and deleting it from the info cache [ 1107.201195] env[62692]: DEBUG nova.network.neutron [req-a49e6360-af9a-435b-bc3d-e9956c47cf85 req-d9c0e90e-c21a-4f53-8b1e-51e6f4ed158f service nova] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1107.483101] env[62692]: DEBUG oslo_vmware.api [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141846, 'name': ReconfigVM_Task, 'duration_secs': 0.625139} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.484540] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Reconfigured VM instance instance-00000067 to attach disk [datastore2] adfb0194-e94b-4a87-8554-a5c262fe9a70/adfb0194-e94b-4a87-8554-a5c262fe9a70.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1107.486451] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-84a8cc97-ea03-4e81-a99c-696bef64882c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.493509] env[62692]: DEBUG oslo_vmware.api [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1107.493509] env[62692]: value = "task-1141847" [ 1107.493509] env[62692]: _type = "Task" [ 1107.493509] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.506071] env[62692]: DEBUG oslo_vmware.api [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141847, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.552068] env[62692]: DEBUG oslo_concurrency.lockutils [None req-19fad0d0-a67d-44f8-afa1-6fd01c7a30df tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "1d9271aa-d41d-48cd-a879-3306e7212197" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.818s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.565128] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d5d121df-64dc-4db3-b700-b1b98f7ec74f tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "1904edb2-fd72-43b6-9219-8043c5abff0c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.896s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.662203] env[62692]: DEBUG nova.network.neutron [-] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1107.708026] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fdf849da-dbc9-4d4a-91ee-af74614fec20 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.715702] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca7ab272-c2c3-4a97-b60a-a545fb49e1fd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.752637] env[62692]: DEBUG nova.compute.manager [req-a49e6360-af9a-435b-bc3d-e9956c47cf85 req-d9c0e90e-c21a-4f53-8b1e-51e6f4ed158f service nova] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Detach interface failed, port_id=877ab637-ea08-499f-9d34-88ca15e1ed1a, reason: Instance e8f645da-3b4a-4f21-9a17-aebdc8d112df could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1107.819971] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquiring lock "dd8a0d53-5dc8-4b66-add7-f244e600468d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.820428] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Lock "dd8a0d53-5dc8-4b66-add7-f244e600468d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.009590] env[62692]: DEBUG oslo_vmware.api [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141847, 'name': Rename_Task, 'duration_secs': 0.194623} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.010100] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1108.010188] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6d8f0166-5071-4dc6-b92f-45a3e3f7eb71 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.016741] env[62692]: DEBUG oslo_vmware.api [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1108.016741] env[62692]: value = "task-1141848" [ 1108.016741] env[62692]: _type = "Task" [ 1108.016741] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.024839] env[62692]: DEBUG oslo_vmware.api [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141848, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.163596] env[62692]: INFO nova.compute.manager [-] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Took 2.14 seconds to deallocate network for instance. [ 1108.322403] env[62692]: DEBUG nova.compute.manager [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1108.344181] env[62692]: DEBUG nova.compute.manager [req-308c3d56-7884-4709-b2d2-dffffaf0c2b2 req-4ee008f9-78df-4544-a447-37f75c0b9f06 service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Received event network-changed-d51ffd3f-c023-45e2-8bfa-1424b7ccb188 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1108.344181] env[62692]: DEBUG nova.compute.manager [req-308c3d56-7884-4709-b2d2-dffffaf0c2b2 req-4ee008f9-78df-4544-a447-37f75c0b9f06 service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Refreshing instance network info cache due to event network-changed-d51ffd3f-c023-45e2-8bfa-1424b7ccb188. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1108.344181] env[62692]: DEBUG oslo_concurrency.lockutils [req-308c3d56-7884-4709-b2d2-dffffaf0c2b2 req-4ee008f9-78df-4544-a447-37f75c0b9f06 service nova] Acquiring lock "refresh_cache-94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1108.344181] env[62692]: DEBUG oslo_concurrency.lockutils [req-308c3d56-7884-4709-b2d2-dffffaf0c2b2 req-4ee008f9-78df-4544-a447-37f75c0b9f06 service nova] Acquired lock "refresh_cache-94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.344181] env[62692]: DEBUG nova.network.neutron [req-308c3d56-7884-4709-b2d2-dffffaf0c2b2 req-4ee008f9-78df-4544-a447-37f75c0b9f06 service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Refreshing network info cache for port d51ffd3f-c023-45e2-8bfa-1424b7ccb188 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1108.534294] env[62692]: DEBUG oslo_vmware.api [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141848, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.672250] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.673065] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.673065] env[62692]: DEBUG nova.objects.instance [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lazy-loading 'resources' on Instance uuid e8f645da-3b4a-4f21-9a17-aebdc8d112df {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1108.858416] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.964546] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "e55a1a5b-3656-48d7-a610-04610b43999a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.964784] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "e55a1a5b-3656-48d7-a610-04610b43999a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.029461] env[62692]: DEBUG oslo_vmware.api [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141848, 'name': PowerOnVM_Task, 'duration_secs': 0.909352} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.029855] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1109.030115] env[62692]: INFO nova.compute.manager [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Took 9.18 seconds to spawn the instance on the hypervisor. [ 1109.030585] env[62692]: DEBUG nova.compute.manager [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1109.031733] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-111294a7-a7b0-4a58-8251-59a1d832c199 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.226953] env[62692]: DEBUG nova.network.neutron [req-308c3d56-7884-4709-b2d2-dffffaf0c2b2 req-4ee008f9-78df-4544-a447-37f75c0b9f06 service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Updated VIF entry in instance network info cache for port d51ffd3f-c023-45e2-8bfa-1424b7ccb188. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1109.226953] env[62692]: DEBUG nova.network.neutron [req-308c3d56-7884-4709-b2d2-dffffaf0c2b2 req-4ee008f9-78df-4544-a447-37f75c0b9f06 service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Updating instance_info_cache with network_info: [{"id": "d51ffd3f-c023-45e2-8bfa-1424b7ccb188", "address": "fa:16:3e:a8:44:f1", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd51ffd3f-c0", "ovs_interfaceid": "d51ffd3f-c023-45e2-8bfa-1424b7ccb188", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1109.392038] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21f2ac5f-4652-4744-99e1-a3c815743c68 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.403469] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05a685c4-e3dc-40c6-95ec-9d40a1a1b618 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.440055] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8995c954-296b-4688-892e-6946c6b98173 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.449046] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74313676-1d4c-4fd5-82ad-65ba3abdcd5e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.464768] env[62692]: DEBUG nova.compute.provider_tree [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1109.467598] env[62692]: DEBUG nova.compute.manager [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1109.548818] env[62692]: INFO nova.compute.manager [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Took 16.32 seconds to build instance. [ 1109.727908] env[62692]: DEBUG oslo_concurrency.lockutils [req-308c3d56-7884-4709-b2d2-dffffaf0c2b2 req-4ee008f9-78df-4544-a447-37f75c0b9f06 service nova] Releasing lock "refresh_cache-94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1109.971236] env[62692]: DEBUG nova.scheduler.client.report [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1110.009709] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.051411] env[62692]: DEBUG oslo_concurrency.lockutils [None req-82f556cf-8e62-416d-b3fb-482233bfad85 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "adfb0194-e94b-4a87-8554-a5c262fe9a70" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.826s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.385043] env[62692]: DEBUG nova.compute.manager [req-e9f4fd8a-f4dd-4644-9440-1eac5d49eb4a req-89659687-f9d7-44ff-a841-1b86b064b6dd service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Received event network-changed-a668d6b6-4111-4a94-8bc8-09535b29f895 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1110.385043] env[62692]: DEBUG nova.compute.manager [req-e9f4fd8a-f4dd-4644-9440-1eac5d49eb4a req-89659687-f9d7-44ff-a841-1b86b064b6dd service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Refreshing instance network info cache due to event network-changed-a668d6b6-4111-4a94-8bc8-09535b29f895. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1110.385043] env[62692]: DEBUG oslo_concurrency.lockutils [req-e9f4fd8a-f4dd-4644-9440-1eac5d49eb4a req-89659687-f9d7-44ff-a841-1b86b064b6dd service nova] Acquiring lock "refresh_cache-1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.385043] env[62692]: DEBUG oslo_concurrency.lockutils [req-e9f4fd8a-f4dd-4644-9440-1eac5d49eb4a req-89659687-f9d7-44ff-a841-1b86b064b6dd service nova] Acquired lock "refresh_cache-1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.385043] env[62692]: DEBUG nova.network.neutron [req-e9f4fd8a-f4dd-4644-9440-1eac5d49eb4a req-89659687-f9d7-44ff-a841-1b86b064b6dd service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Refreshing network info cache for port a668d6b6-4111-4a94-8bc8-09535b29f895 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1110.483178] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.810s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.487261] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.627s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.489400] env[62692]: INFO nova.compute.claims [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1110.511628] env[62692]: INFO nova.scheduler.client.report [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Deleted allocations for instance e8f645da-3b4a-4f21-9a17-aebdc8d112df [ 1111.026949] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d760f364-4adc-4e5a-a422-25f9d988b91d tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "e8f645da-3b4a-4f21-9a17-aebdc8d112df" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.150s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.276604] env[62692]: DEBUG oslo_concurrency.lockutils [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "adfb0194-e94b-4a87-8554-a5c262fe9a70" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.276604] env[62692]: DEBUG oslo_concurrency.lockutils [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "adfb0194-e94b-4a87-8554-a5c262fe9a70" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.276604] env[62692]: DEBUG oslo_concurrency.lockutils [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "adfb0194-e94b-4a87-8554-a5c262fe9a70-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.277058] env[62692]: DEBUG oslo_concurrency.lockutils [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "adfb0194-e94b-4a87-8554-a5c262fe9a70-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.277058] env[62692]: DEBUG oslo_concurrency.lockutils [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "adfb0194-e94b-4a87-8554-a5c262fe9a70-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.281115] env[62692]: INFO nova.compute.manager [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Terminating instance [ 1111.284910] env[62692]: DEBUG nova.compute.manager [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1111.284910] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1111.284983] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d08fa1cb-6632-4abc-8990-c259869ed33f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.293950] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1111.294240] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-31864217-d0db-4307-ab88-142ed4f85660 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.302692] env[62692]: DEBUG oslo_vmware.api [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1111.302692] env[62692]: value = "task-1141849" [ 1111.302692] env[62692]: _type = "Task" [ 1111.302692] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.314959] env[62692]: DEBUG oslo_vmware.api [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141849, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.470712] env[62692]: DEBUG nova.network.neutron [req-e9f4fd8a-f4dd-4644-9440-1eac5d49eb4a req-89659687-f9d7-44ff-a841-1b86b064b6dd service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Updated VIF entry in instance network info cache for port a668d6b6-4111-4a94-8bc8-09535b29f895. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1111.471151] env[62692]: DEBUG nova.network.neutron [req-e9f4fd8a-f4dd-4644-9440-1eac5d49eb4a req-89659687-f9d7-44ff-a841-1b86b064b6dd service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Updating instance_info_cache with network_info: [{"id": "a668d6b6-4111-4a94-8bc8-09535b29f895", "address": "fa:16:3e:02:7f:2a", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa668d6b6-41", "ovs_interfaceid": "a668d6b6-4111-4a94-8bc8-09535b29f895", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.663876] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67b59e57-2400-405f-b881-fb041d936180 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.672072] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7932a1e6-ce7a-4ba6-9da5-dd9b835b5fa1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.712054] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac2b4f72-c95f-4194-9d39-563232bba34d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.721283] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cea4d3ce-9603-4a82-bdef-f4a5685c205c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.735559] env[62692]: DEBUG nova.compute.provider_tree [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1111.814151] env[62692]: DEBUG oslo_vmware.api [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141849, 'name': PowerOffVM_Task, 'duration_secs': 0.210181} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.814438] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1111.814614] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1111.814869] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8e5b074f-6ca6-4784-8777-9b375b2fd397 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.879278] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1111.879535] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1111.879727] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Deleting the datastore file [datastore2] adfb0194-e94b-4a87-8554-a5c262fe9a70 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1111.879997] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eb41a458-cb22-49bb-893d-ab4c8a6f7292 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.887889] env[62692]: DEBUG oslo_vmware.api [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for the task: (returnval){ [ 1111.887889] env[62692]: value = "task-1141851" [ 1111.887889] env[62692]: _type = "Task" [ 1111.887889] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.897184] env[62692]: DEBUG oslo_vmware.api [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141851, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.975158] env[62692]: DEBUG nova.compute.manager [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1111.975977] env[62692]: DEBUG oslo_concurrency.lockutils [req-e9f4fd8a-f4dd-4644-9440-1eac5d49eb4a req-89659687-f9d7-44ff-a841-1b86b064b6dd service nova] Releasing lock "refresh_cache-1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.975977] env[62692]: DEBUG nova.compute.manager [req-e9f4fd8a-f4dd-4644-9440-1eac5d49eb4a req-89659687-f9d7-44ff-a841-1b86b064b6dd service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Received event network-changed-a668d6b6-4111-4a94-8bc8-09535b29f895 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1111.976121] env[62692]: DEBUG nova.compute.manager [req-e9f4fd8a-f4dd-4644-9440-1eac5d49eb4a req-89659687-f9d7-44ff-a841-1b86b064b6dd service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Refreshing instance network info cache due to event network-changed-a668d6b6-4111-4a94-8bc8-09535b29f895. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1111.976379] env[62692]: DEBUG oslo_concurrency.lockutils [req-e9f4fd8a-f4dd-4644-9440-1eac5d49eb4a req-89659687-f9d7-44ff-a841-1b86b064b6dd service nova] Acquiring lock "refresh_cache-1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.976471] env[62692]: DEBUG oslo_concurrency.lockutils [req-e9f4fd8a-f4dd-4644-9440-1eac5d49eb4a req-89659687-f9d7-44ff-a841-1b86b064b6dd service nova] Acquired lock "refresh_cache-1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.976728] env[62692]: DEBUG nova.network.neutron [req-e9f4fd8a-f4dd-4644-9440-1eac5d49eb4a req-89659687-f9d7-44ff-a841-1b86b064b6dd service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Refreshing network info cache for port a668d6b6-4111-4a94-8bc8-09535b29f895 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1111.978545] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb8aa248-962f-4aea-b3d8-2688b3463f3e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.238855] env[62692]: DEBUG nova.scheduler.client.report [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1112.397865] env[62692]: DEBUG oslo_vmware.api [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Task: {'id': task-1141851, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162386} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.398198] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1112.398366] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1112.398625] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1112.398742] env[62692]: INFO nova.compute.manager [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1112.398959] env[62692]: DEBUG oslo.service.loopingcall [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1112.399222] env[62692]: DEBUG nova.compute.manager [-] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1112.399328] env[62692]: DEBUG nova.network.neutron [-] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1112.493750] env[62692]: INFO nova.compute.manager [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] instance snapshotting [ 1112.493750] env[62692]: DEBUG nova.objects.instance [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lazy-loading 'flavor' on Instance uuid 90932f05-fad3-4c6d-87ae-ab059351b0be {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1112.744193] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.258s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.744818] env[62692]: DEBUG nova.compute.manager [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1112.753704] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.744s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1112.755213] env[62692]: INFO nova.compute.claims [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1112.850730] env[62692]: DEBUG nova.network.neutron [req-e9f4fd8a-f4dd-4644-9440-1eac5d49eb4a req-89659687-f9d7-44ff-a841-1b86b064b6dd service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Updated VIF entry in instance network info cache for port a668d6b6-4111-4a94-8bc8-09535b29f895. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1112.850730] env[62692]: DEBUG nova.network.neutron [req-e9f4fd8a-f4dd-4644-9440-1eac5d49eb4a req-89659687-f9d7-44ff-a841-1b86b064b6dd service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Updating instance_info_cache with network_info: [{"id": "a668d6b6-4111-4a94-8bc8-09535b29f895", "address": "fa:16:3e:02:7f:2a", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa668d6b6-41", "ovs_interfaceid": "a668d6b6-4111-4a94-8bc8-09535b29f895", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1112.980448] env[62692]: DEBUG nova.compute.manager [req-b39a50e0-91ac-4b0a-8723-c6610543c42d req-32d3ae7e-3cf7-46e8-a935-9a7770beaa73 service nova] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Received event network-vif-deleted-da5f5c81-a3f5-46d7-b2eb-364476be8845 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1112.980448] env[62692]: INFO nova.compute.manager [req-b39a50e0-91ac-4b0a-8723-c6610543c42d req-32d3ae7e-3cf7-46e8-a935-9a7770beaa73 service nova] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Neutron deleted interface da5f5c81-a3f5-46d7-b2eb-364476be8845; detaching it from the instance and deleting it from the info cache [ 1112.980448] env[62692]: DEBUG nova.network.neutron [req-b39a50e0-91ac-4b0a-8723-c6610543c42d req-32d3ae7e-3cf7-46e8-a935-9a7770beaa73 service nova] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1112.999561] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb3d0614-4806-4190-bcf3-967e6191bf85 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.020253] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cebc9ba-e546-46a4-9264-108b4d126f01 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.224636] env[62692]: DEBUG nova.network.neutron [-] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1113.254970] env[62692]: DEBUG nova.compute.utils [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1113.256340] env[62692]: DEBUG nova.compute.manager [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1113.256696] env[62692]: DEBUG nova.network.neutron [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1113.299526] env[62692]: DEBUG nova.policy [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6993a074d1384d1e8912e056f0a4caef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b1c44eb72b334ee3add29026d312e37b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1113.353433] env[62692]: DEBUG oslo_concurrency.lockutils [req-e9f4fd8a-f4dd-4644-9440-1eac5d49eb4a req-89659687-f9d7-44ff-a841-1b86b064b6dd service nova] Releasing lock "refresh_cache-1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1113.353433] env[62692]: DEBUG nova.compute.manager [req-e9f4fd8a-f4dd-4644-9440-1eac5d49eb4a req-89659687-f9d7-44ff-a841-1b86b064b6dd service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Received event network-changed-d51ffd3f-c023-45e2-8bfa-1424b7ccb188 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1113.353598] env[62692]: DEBUG nova.compute.manager [req-e9f4fd8a-f4dd-4644-9440-1eac5d49eb4a req-89659687-f9d7-44ff-a841-1b86b064b6dd service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Refreshing instance network info cache due to event network-changed-d51ffd3f-c023-45e2-8bfa-1424b7ccb188. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1113.353787] env[62692]: DEBUG oslo_concurrency.lockutils [req-e9f4fd8a-f4dd-4644-9440-1eac5d49eb4a req-89659687-f9d7-44ff-a841-1b86b064b6dd service nova] Acquiring lock "refresh_cache-94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1113.353937] env[62692]: DEBUG oslo_concurrency.lockutils [req-e9f4fd8a-f4dd-4644-9440-1eac5d49eb4a req-89659687-f9d7-44ff-a841-1b86b064b6dd service nova] Acquired lock "refresh_cache-94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1113.354194] env[62692]: DEBUG nova.network.neutron [req-e9f4fd8a-f4dd-4644-9440-1eac5d49eb4a req-89659687-f9d7-44ff-a841-1b86b064b6dd service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Refreshing network info cache for port d51ffd3f-c023-45e2-8bfa-1424b7ccb188 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1113.483683] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-08ad37ca-62e1-4589-83f3-f3bce68cda99 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.495530] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6015a375-3d9b-43eb-81a1-8f52c1410edc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.511374] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "c2f2a485-363e-41af-9b3f-37092947ca08" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.511622] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "c2f2a485-363e-41af-9b3f-37092947ca08" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.527067] env[62692]: DEBUG nova.compute.manager [req-b39a50e0-91ac-4b0a-8723-c6610543c42d req-32d3ae7e-3cf7-46e8-a935-9a7770beaa73 service nova] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Detach interface failed, port_id=da5f5c81-a3f5-46d7-b2eb-364476be8845, reason: Instance adfb0194-e94b-4a87-8554-a5c262fe9a70 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1113.531487] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Creating Snapshot of the VM instance {{(pid=62692) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1113.531835] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8b0a22a5-e409-4bdb-b8db-9d02bc514ef7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.536937] env[62692]: DEBUG nova.network.neutron [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Successfully created port: f8a6a1d2-a31b-451a-854c-8efc6e7aa879 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1113.540971] env[62692]: DEBUG oslo_vmware.api [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1113.540971] env[62692]: value = "task-1141852" [ 1113.540971] env[62692]: _type = "Task" [ 1113.540971] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.550385] env[62692]: DEBUG oslo_vmware.api [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141852, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.597715] env[62692]: DEBUG oslo_concurrency.lockutils [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "1c7ce408-6e9c-43bd-8d81-6acb12353ed8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.598097] env[62692]: DEBUG oslo_concurrency.lockutils [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "1c7ce408-6e9c-43bd-8d81-6acb12353ed8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.727160] env[62692]: INFO nova.compute.manager [-] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Took 1.33 seconds to deallocate network for instance. [ 1113.760750] env[62692]: DEBUG nova.compute.manager [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1113.969138] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc9a796b-74f6-4dd1-86e6-1d08495e7a72 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.978307] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b89e6915-075b-4460-93ac-89b82fdcbc0e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.014827] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e73fa9ef-7149-4ce1-805f-842aca145f1a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.017478] env[62692]: DEBUG nova.compute.manager [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1114.027055] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35e7956c-028a-483d-92fe-520fe64f2559 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.043920] env[62692]: DEBUG nova.compute.provider_tree [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1114.054277] env[62692]: DEBUG oslo_vmware.api [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141852, 'name': CreateSnapshot_Task, 'duration_secs': 0.514211} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.057072] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Created Snapshot of the VM instance {{(pid=62692) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1114.058502] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ebd01aa-0c11-4132-842e-487cdd23932e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.100138] env[62692]: DEBUG nova.compute.manager [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1114.143128] env[62692]: DEBUG nova.network.neutron [req-e9f4fd8a-f4dd-4644-9440-1eac5d49eb4a req-89659687-f9d7-44ff-a841-1b86b064b6dd service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Updated VIF entry in instance network info cache for port d51ffd3f-c023-45e2-8bfa-1424b7ccb188. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1114.143532] env[62692]: DEBUG nova.network.neutron [req-e9f4fd8a-f4dd-4644-9440-1eac5d49eb4a req-89659687-f9d7-44ff-a841-1b86b064b6dd service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Updating instance_info_cache with network_info: [{"id": "d51ffd3f-c023-45e2-8bfa-1424b7ccb188", "address": "fa:16:3e:a8:44:f1", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd51ffd3f-c0", "ovs_interfaceid": "d51ffd3f-c023-45e2-8bfa-1424b7ccb188", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1114.233823] env[62692]: DEBUG oslo_concurrency.lockutils [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.536871] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.550604] env[62692]: DEBUG nova.scheduler.client.report [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1114.577631] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Creating linked-clone VM from snapshot {{(pid=62692) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1114.578254] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-4310f2ce-f4f1-4212-96db-3740c8136d02 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.588492] env[62692]: DEBUG oslo_vmware.api [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1114.588492] env[62692]: value = "task-1141853" [ 1114.588492] env[62692]: _type = "Task" [ 1114.588492] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.596853] env[62692]: DEBUG oslo_vmware.api [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141853, 'name': CloneVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.617869] env[62692]: DEBUG oslo_concurrency.lockutils [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.646434] env[62692]: DEBUG oslo_concurrency.lockutils [req-e9f4fd8a-f4dd-4644-9440-1eac5d49eb4a req-89659687-f9d7-44ff-a841-1b86b064b6dd service nova] Releasing lock "refresh_cache-94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1114.773073] env[62692]: DEBUG nova.compute.manager [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1114.796083] env[62692]: DEBUG nova.virt.hardware [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1114.796389] env[62692]: DEBUG nova.virt.hardware [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1114.796563] env[62692]: DEBUG nova.virt.hardware [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1114.796756] env[62692]: DEBUG nova.virt.hardware [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1114.796906] env[62692]: DEBUG nova.virt.hardware [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1114.797071] env[62692]: DEBUG nova.virt.hardware [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1114.797291] env[62692]: DEBUG nova.virt.hardware [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1114.797460] env[62692]: DEBUG nova.virt.hardware [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1114.797631] env[62692]: DEBUG nova.virt.hardware [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1114.797798] env[62692]: DEBUG nova.virt.hardware [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1114.797975] env[62692]: DEBUG nova.virt.hardware [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1114.798855] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9df6bebc-9d53-4acd-980b-11894d2b146f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.808821] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-026e0442-859c-46cf-a508-24a4c955f593 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.012115] env[62692]: DEBUG nova.compute.manager [req-71c8b663-e543-4700-871e-3e478c0ee6cf req-2845134f-1375-47ac-85dd-3693a31d4d8e service nova] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Received event network-vif-plugged-f8a6a1d2-a31b-451a-854c-8efc6e7aa879 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1115.012115] env[62692]: DEBUG oslo_concurrency.lockutils [req-71c8b663-e543-4700-871e-3e478c0ee6cf req-2845134f-1375-47ac-85dd-3693a31d4d8e service nova] Acquiring lock "dd8a0d53-5dc8-4b66-add7-f244e600468d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.012115] env[62692]: DEBUG oslo_concurrency.lockutils [req-71c8b663-e543-4700-871e-3e478c0ee6cf req-2845134f-1375-47ac-85dd-3693a31d4d8e service nova] Lock "dd8a0d53-5dc8-4b66-add7-f244e600468d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.012115] env[62692]: DEBUG oslo_concurrency.lockutils [req-71c8b663-e543-4700-871e-3e478c0ee6cf req-2845134f-1375-47ac-85dd-3693a31d4d8e service nova] Lock "dd8a0d53-5dc8-4b66-add7-f244e600468d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.012115] env[62692]: DEBUG nova.compute.manager [req-71c8b663-e543-4700-871e-3e478c0ee6cf req-2845134f-1375-47ac-85dd-3693a31d4d8e service nova] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] No waiting events found dispatching network-vif-plugged-f8a6a1d2-a31b-451a-854c-8efc6e7aa879 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1115.012573] env[62692]: WARNING nova.compute.manager [req-71c8b663-e543-4700-871e-3e478c0ee6cf req-2845134f-1375-47ac-85dd-3693a31d4d8e service nova] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Received unexpected event network-vif-plugged-f8a6a1d2-a31b-451a-854c-8efc6e7aa879 for instance with vm_state building and task_state spawning. [ 1115.055970] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.302s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.055970] env[62692]: DEBUG nova.compute.manager [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1115.058772] env[62692]: DEBUG oslo_concurrency.lockutils [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.825s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.059020] env[62692]: DEBUG nova.objects.instance [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lazy-loading 'resources' on Instance uuid adfb0194-e94b-4a87-8554-a5c262fe9a70 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1115.101024] env[62692]: DEBUG oslo_vmware.api [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141853, 'name': CloneVM_Task} progress is 94%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.101792] env[62692]: DEBUG nova.network.neutron [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Successfully updated port: f8a6a1d2-a31b-451a-854c-8efc6e7aa879 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1115.561725] env[62692]: DEBUG nova.compute.utils [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1115.566210] env[62692]: DEBUG nova.compute.manager [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1115.566391] env[62692]: DEBUG nova.network.neutron [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1116.282206] env[62692]: DEBUG nova.policy [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a28f5e0e0af443c5be82214d446c221d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '589895ae20394e2fa525707bf7d09f2a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1116.286066] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquiring lock "refresh_cache-dd8a0d53-5dc8-4b66-add7-f244e600468d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1116.286207] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquired lock "refresh_cache-dd8a0d53-5dc8-4b66-add7-f244e600468d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.286348] env[62692]: DEBUG nova.network.neutron [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1116.287757] env[62692]: DEBUG nova.compute.manager [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1116.301197] env[62692]: DEBUG oslo_vmware.api [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141853, 'name': CloneVM_Task, 'duration_secs': 1.394006} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.301197] env[62692]: INFO nova.virt.vmwareapi.vmops [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Created linked-clone VM from snapshot [ 1116.304467] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-121f5fff-1164-48bc-a04a-31b5ef6d6016 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.316625] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Uploading image b7b0878a-b2c2-4474-9dbb-47b954aa8a86 {{(pid=62692) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1116.342168] env[62692]: DEBUG oslo_vmware.rw_handles [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1116.342168] env[62692]: value = "vm-249054" [ 1116.342168] env[62692]: _type = "VirtualMachine" [ 1116.342168] env[62692]: }. {{(pid=62692) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1116.342510] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-e6139af2-ebdc-4290-b26d-e3bc60921766 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.354191] env[62692]: DEBUG oslo_vmware.rw_handles [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lease: (returnval){ [ 1116.354191] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52fcfa2f-602d-9b3e-0793-c18a5585e0cb" [ 1116.354191] env[62692]: _type = "HttpNfcLease" [ 1116.354191] env[62692]: } obtained for exporting VM: (result){ [ 1116.354191] env[62692]: value = "vm-249054" [ 1116.354191] env[62692]: _type = "VirtualMachine" [ 1116.354191] env[62692]: }. {{(pid=62692) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1116.354191] env[62692]: DEBUG oslo_vmware.api [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the lease: (returnval){ [ 1116.354191] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52fcfa2f-602d-9b3e-0793-c18a5585e0cb" [ 1116.354191] env[62692]: _type = "HttpNfcLease" [ 1116.354191] env[62692]: } to be ready. {{(pid=62692) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1116.359954] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1116.359954] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52fcfa2f-602d-9b3e-0793-c18a5585e0cb" [ 1116.359954] env[62692]: _type = "HttpNfcLease" [ 1116.359954] env[62692]: } is initializing. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1116.445850] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9259b286-4391-42bd-a14a-3fd6343783e4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.454749] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3226cc67-f0d0-4941-bce0-1cd98ba21afe {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.485576] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fdb5499-ce4c-43b2-9001-af73c611a50d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.490947] env[62692]: DEBUG oslo_concurrency.lockutils [None req-32cc40cc-426b-46bb-9325-d895f694d4fc tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.491232] env[62692]: DEBUG oslo_concurrency.lockutils [None req-32cc40cc-426b-46bb-9325-d895f694d4fc tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.496513] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ecfa2a1-0226-43f6-8b7e-861f0a225030 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.513356] env[62692]: DEBUG nova.compute.provider_tree [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1116.584830] env[62692]: DEBUG nova.network.neutron [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Successfully created port: 5e50f271-31a9-4755-a612-6b971179167a {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1116.828423] env[62692]: DEBUG nova.network.neutron [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1116.861108] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1116.861108] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52fcfa2f-602d-9b3e-0793-c18a5585e0cb" [ 1116.861108] env[62692]: _type = "HttpNfcLease" [ 1116.861108] env[62692]: } is ready. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1116.861108] env[62692]: DEBUG oslo_vmware.rw_handles [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1116.861108] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52fcfa2f-602d-9b3e-0793-c18a5585e0cb" [ 1116.861108] env[62692]: _type = "HttpNfcLease" [ 1116.861108] env[62692]: }. {{(pid=62692) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1116.861910] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d08a241-f95e-4a0a-8c94-7ecfc56e4f34 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.869541] env[62692]: DEBUG oslo_vmware.rw_handles [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ea7a3a-3def-d36e-02ce-55b9740d4384/disk-0.vmdk from lease info. {{(pid=62692) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1116.869728] env[62692]: DEBUG oslo_vmware.rw_handles [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ea7a3a-3def-d36e-02ce-55b9740d4384/disk-0.vmdk for reading. {{(pid=62692) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1116.961249] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-45a0b6a5-d736-4832-95d2-d712e8dcc470 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.997042] env[62692]: DEBUG nova.compute.utils [None req-32cc40cc-426b-46bb-9325-d895f694d4fc tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1117.004649] env[62692]: DEBUG nova.network.neutron [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Updating instance_info_cache with network_info: [{"id": "f8a6a1d2-a31b-451a-854c-8efc6e7aa879", "address": "fa:16:3e:9f:bb:e7", "network": {"id": "50c1330e-1a5f-42d1-9987-79ddcc0ad7aa", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-813901268-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b1c44eb72b334ee3add29026d312e37b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f92f0b92-d6fb-4d00-8ad5-6b3809ed5493", "external-id": "nsx-vlan-transportzone-48", "segmentation_id": 48, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8a6a1d2-a3", "ovs_interfaceid": "f8a6a1d2-a31b-451a-854c-8efc6e7aa879", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1117.016024] env[62692]: DEBUG nova.scheduler.client.report [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1117.037626] env[62692]: DEBUG nova.compute.manager [req-9fd6a766-346c-4701-85cc-9cd0b1bd5754 req-1ab83b56-6d24-4531-b39f-211c15e18473 service nova] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Received event network-changed-f8a6a1d2-a31b-451a-854c-8efc6e7aa879 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1117.037822] env[62692]: DEBUG nova.compute.manager [req-9fd6a766-346c-4701-85cc-9cd0b1bd5754 req-1ab83b56-6d24-4531-b39f-211c15e18473 service nova] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Refreshing instance network info cache due to event network-changed-f8a6a1d2-a31b-451a-854c-8efc6e7aa879. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1117.038010] env[62692]: DEBUG oslo_concurrency.lockutils [req-9fd6a766-346c-4701-85cc-9cd0b1bd5754 req-1ab83b56-6d24-4531-b39f-211c15e18473 service nova] Acquiring lock "refresh_cache-dd8a0d53-5dc8-4b66-add7-f244e600468d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1117.301498] env[62692]: DEBUG nova.compute.manager [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1117.325112] env[62692]: DEBUG nova.virt.hardware [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1117.325469] env[62692]: DEBUG nova.virt.hardware [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1117.325700] env[62692]: DEBUG nova.virt.hardware [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1117.326008] env[62692]: DEBUG nova.virt.hardware [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1117.326174] env[62692]: DEBUG nova.virt.hardware [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1117.326334] env[62692]: DEBUG nova.virt.hardware [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1117.326550] env[62692]: DEBUG nova.virt.hardware [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1117.326758] env[62692]: DEBUG nova.virt.hardware [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1117.327093] env[62692]: DEBUG nova.virt.hardware [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1117.327288] env[62692]: DEBUG nova.virt.hardware [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1117.327684] env[62692]: DEBUG nova.virt.hardware [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1117.328601] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e3e41d-c572-4d47-b48b-c6fedea4bc6d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.337315] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b4eaa1b-2ba5-48bd-941c-137431206984 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.501044] env[62692]: DEBUG oslo_concurrency.lockutils [None req-32cc40cc-426b-46bb-9325-d895f694d4fc tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.507314] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Releasing lock "refresh_cache-dd8a0d53-5dc8-4b66-add7-f244e600468d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1117.507626] env[62692]: DEBUG nova.compute.manager [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Instance network_info: |[{"id": "f8a6a1d2-a31b-451a-854c-8efc6e7aa879", "address": "fa:16:3e:9f:bb:e7", "network": {"id": "50c1330e-1a5f-42d1-9987-79ddcc0ad7aa", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-813901268-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b1c44eb72b334ee3add29026d312e37b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f92f0b92-d6fb-4d00-8ad5-6b3809ed5493", "external-id": "nsx-vlan-transportzone-48", "segmentation_id": 48, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8a6a1d2-a3", "ovs_interfaceid": "f8a6a1d2-a31b-451a-854c-8efc6e7aa879", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1117.507910] env[62692]: DEBUG oslo_concurrency.lockutils [req-9fd6a766-346c-4701-85cc-9cd0b1bd5754 req-1ab83b56-6d24-4531-b39f-211c15e18473 service nova] Acquired lock "refresh_cache-dd8a0d53-5dc8-4b66-add7-f244e600468d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1117.508105] env[62692]: DEBUG nova.network.neutron [req-9fd6a766-346c-4701-85cc-9cd0b1bd5754 req-1ab83b56-6d24-4531-b39f-211c15e18473 service nova] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Refreshing network info cache for port f8a6a1d2-a31b-451a-854c-8efc6e7aa879 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1117.509621] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9f:bb:e7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f92f0b92-d6fb-4d00-8ad5-6b3809ed5493', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f8a6a1d2-a31b-451a-854c-8efc6e7aa879', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1117.518811] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Creating folder: Project (b1c44eb72b334ee3add29026d312e37b). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1117.519966] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bd03941b-a703-407a-8d9d-f49b4c9d6e00 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.522813] env[62692]: DEBUG oslo_concurrency.lockutils [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.464s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.525362] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.989s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.528342] env[62692]: INFO nova.compute.claims [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1117.543221] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Created folder: Project (b1c44eb72b334ee3add29026d312e37b) in parent group-v248868. [ 1117.543588] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Creating folder: Instances. Parent ref: group-v249055. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1117.544285] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-775817e9-f0f2-47a3-8dd9-e1a2c54813d9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.547171] env[62692]: INFO nova.scheduler.client.report [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Deleted allocations for instance adfb0194-e94b-4a87-8554-a5c262fe9a70 [ 1117.558422] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Created folder: Instances in parent group-v249055. [ 1117.559017] env[62692]: DEBUG oslo.service.loopingcall [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1117.559387] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1117.559704] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e243e67a-be39-4b44-bf10-47a50b0d6186 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.584480] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1117.584480] env[62692]: value = "task-1141857" [ 1117.584480] env[62692]: _type = "Task" [ 1117.584480] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.594520] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141857, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.055548] env[62692]: DEBUG oslo_concurrency.lockutils [None req-889ad64f-a181-4e24-8320-e94ca2a5fb08 tempest-ServerDiskConfigTestJSON-102120146 tempest-ServerDiskConfigTestJSON-102120146-project-member] Lock "adfb0194-e94b-4a87-8554-a5c262fe9a70" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.779s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.081802] env[62692]: DEBUG nova.network.neutron [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Successfully updated port: 5e50f271-31a9-4755-a612-6b971179167a {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1118.098749] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141857, 'name': CreateVM_Task, 'duration_secs': 0.44203} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.099645] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1118.100614] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.100986] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.101591] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1118.101900] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9b5479d-040a-41b8-bbac-74178416f83a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.109187] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1118.109187] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52595cac-747e-44fb-e09c-f6bb84bf837a" [ 1118.109187] env[62692]: _type = "Task" [ 1118.109187] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.118830] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52595cac-747e-44fb-e09c-f6bb84bf837a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.244258] env[62692]: DEBUG nova.network.neutron [req-9fd6a766-346c-4701-85cc-9cd0b1bd5754 req-1ab83b56-6d24-4531-b39f-211c15e18473 service nova] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Updated VIF entry in instance network info cache for port f8a6a1d2-a31b-451a-854c-8efc6e7aa879. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1118.244785] env[62692]: DEBUG nova.network.neutron [req-9fd6a766-346c-4701-85cc-9cd0b1bd5754 req-1ab83b56-6d24-4531-b39f-211c15e18473 service nova] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Updating instance_info_cache with network_info: [{"id": "f8a6a1d2-a31b-451a-854c-8efc6e7aa879", "address": "fa:16:3e:9f:bb:e7", "network": {"id": "50c1330e-1a5f-42d1-9987-79ddcc0ad7aa", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-813901268-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b1c44eb72b334ee3add29026d312e37b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f92f0b92-d6fb-4d00-8ad5-6b3809ed5493", "external-id": "nsx-vlan-transportzone-48", "segmentation_id": 48, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8a6a1d2-a3", "ovs_interfaceid": "f8a6a1d2-a31b-451a-854c-8efc6e7aa879", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1118.576398] env[62692]: DEBUG oslo_concurrency.lockutils [None req-32cc40cc-426b-46bb-9325-d895f694d4fc tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.577161] env[62692]: DEBUG oslo_concurrency.lockutils [None req-32cc40cc-426b-46bb-9325-d895f694d4fc tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.577161] env[62692]: INFO nova.compute.manager [None req-32cc40cc-426b-46bb-9325-d895f694d4fc tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Attaching volume 02a5babd-e299-40a6-a55f-e70273241e04 to /dev/sdb [ 1118.585600] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "refresh_cache-e55a1a5b-3656-48d7-a610-04610b43999a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.585836] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired lock "refresh_cache-e55a1a5b-3656-48d7-a610-04610b43999a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.585994] env[62692]: DEBUG nova.network.neutron [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1118.622106] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e080ee6-1491-4494-83d7-f01c57fdb9c7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.628421] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52595cac-747e-44fb-e09c-f6bb84bf837a, 'name': SearchDatastore_Task, 'duration_secs': 0.023776} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.631761] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1118.632019] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1118.632288] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.632443] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.632627] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1118.633092] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-53e0f3ba-915c-450c-94e2-e293d41f2c60 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.637877] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10870811-2023-4b8d-8d1d-b5f8c981c017 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.644121] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1118.644334] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1118.647447] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b52e0f1-00e1-4ad6-9e7f-b3d1e6bc6886 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.654422] env[62692]: DEBUG nova.virt.block_device [None req-32cc40cc-426b-46bb-9325-d895f694d4fc tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Updating existing volume attachment record: 2e462ba8-6401-49fd-8df9-4908fc261b2e {{(pid=62692) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1118.659552] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1118.659552] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5205d9d0-81b6-0b27-4d9a-843958f6be76" [ 1118.659552] env[62692]: _type = "Task" [ 1118.659552] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.670545] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5205d9d0-81b6-0b27-4d9a-843958f6be76, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.718276] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c37f59c-cbab-43e6-868a-405fdff40a19 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.725924] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcb38819-3916-4198-96cc-bd6e3978e0bb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.758896] env[62692]: DEBUG oslo_concurrency.lockutils [req-9fd6a766-346c-4701-85cc-9cd0b1bd5754 req-1ab83b56-6d24-4531-b39f-211c15e18473 service nova] Releasing lock "refresh_cache-dd8a0d53-5dc8-4b66-add7-f244e600468d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1118.760138] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2925d51-6531-4587-bd79-2ef54d2c7c21 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.768738] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96882b60-44ee-48bb-a6b3-a9297a3d14e2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.783200] env[62692]: DEBUG nova.compute.provider_tree [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1119.070394] env[62692]: DEBUG nova.compute.manager [req-7afe8964-f151-4396-a4bf-f3bd26c33f06 req-72f778c6-a36c-4332-b98a-ac53e0ac32d6 service nova] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Received event network-vif-plugged-5e50f271-31a9-4755-a612-6b971179167a {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1119.072748] env[62692]: DEBUG oslo_concurrency.lockutils [req-7afe8964-f151-4396-a4bf-f3bd26c33f06 req-72f778c6-a36c-4332-b98a-ac53e0ac32d6 service nova] Acquiring lock "e55a1a5b-3656-48d7-a610-04610b43999a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.072748] env[62692]: DEBUG oslo_concurrency.lockutils [req-7afe8964-f151-4396-a4bf-f3bd26c33f06 req-72f778c6-a36c-4332-b98a-ac53e0ac32d6 service nova] Lock "e55a1a5b-3656-48d7-a610-04610b43999a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.072748] env[62692]: DEBUG oslo_concurrency.lockutils [req-7afe8964-f151-4396-a4bf-f3bd26c33f06 req-72f778c6-a36c-4332-b98a-ac53e0ac32d6 service nova] Lock "e55a1a5b-3656-48d7-a610-04610b43999a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.072748] env[62692]: DEBUG nova.compute.manager [req-7afe8964-f151-4396-a4bf-f3bd26c33f06 req-72f778c6-a36c-4332-b98a-ac53e0ac32d6 service nova] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] No waiting events found dispatching network-vif-plugged-5e50f271-31a9-4755-a612-6b971179167a {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1119.072748] env[62692]: WARNING nova.compute.manager [req-7afe8964-f151-4396-a4bf-f3bd26c33f06 req-72f778c6-a36c-4332-b98a-ac53e0ac32d6 service nova] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Received unexpected event network-vif-plugged-5e50f271-31a9-4755-a612-6b971179167a for instance with vm_state building and task_state spawning. [ 1119.072748] env[62692]: DEBUG nova.compute.manager [req-7afe8964-f151-4396-a4bf-f3bd26c33f06 req-72f778c6-a36c-4332-b98a-ac53e0ac32d6 service nova] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Received event network-changed-5e50f271-31a9-4755-a612-6b971179167a {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1119.072748] env[62692]: DEBUG nova.compute.manager [req-7afe8964-f151-4396-a4bf-f3bd26c33f06 req-72f778c6-a36c-4332-b98a-ac53e0ac32d6 service nova] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Refreshing instance network info cache due to event network-changed-5e50f271-31a9-4755-a612-6b971179167a. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1119.073538] env[62692]: DEBUG oslo_concurrency.lockutils [req-7afe8964-f151-4396-a4bf-f3bd26c33f06 req-72f778c6-a36c-4332-b98a-ac53e0ac32d6 service nova] Acquiring lock "refresh_cache-e55a1a5b-3656-48d7-a610-04610b43999a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1119.125324] env[62692]: DEBUG nova.network.neutron [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1119.173388] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5205d9d0-81b6-0b27-4d9a-843958f6be76, 'name': SearchDatastore_Task, 'duration_secs': 0.017441} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.174478] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cabf73a4-b652-482d-ad1c-77bf1fdc4fff {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.181690] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1119.181690] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52d39576-32f8-89d0-f9f2-e3718865c973" [ 1119.181690] env[62692]: _type = "Task" [ 1119.181690] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.191664] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52d39576-32f8-89d0-f9f2-e3718865c973, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.276305] env[62692]: DEBUG nova.network.neutron [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Updating instance_info_cache with network_info: [{"id": "5e50f271-31a9-4755-a612-6b971179167a", "address": "fa:16:3e:75:3c:39", "network": {"id": "dd260020-2d74-46ba-aa78-af29bdefa176", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-617179655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "589895ae20394e2fa525707bf7d09f2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e50f271-31", "ovs_interfaceid": "5e50f271-31a9-4755-a612-6b971179167a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1119.286526] env[62692]: DEBUG nova.scheduler.client.report [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1119.694420] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52d39576-32f8-89d0-f9f2-e3718865c973, 'name': SearchDatastore_Task, 'duration_secs': 0.013321} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.694837] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1119.695062] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] dd8a0d53-5dc8-4b66-add7-f244e600468d/dd8a0d53-5dc8-4b66-add7-f244e600468d.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1119.695300] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ad6b88bb-ea82-4410-ba70-2d7cf1c4371d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.703404] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1119.703404] env[62692]: value = "task-1141859" [ 1119.703404] env[62692]: _type = "Task" [ 1119.703404] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.714842] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141859, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.779746] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Releasing lock "refresh_cache-e55a1a5b-3656-48d7-a610-04610b43999a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1119.780180] env[62692]: DEBUG nova.compute.manager [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Instance network_info: |[{"id": "5e50f271-31a9-4755-a612-6b971179167a", "address": "fa:16:3e:75:3c:39", "network": {"id": "dd260020-2d74-46ba-aa78-af29bdefa176", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-617179655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "589895ae20394e2fa525707bf7d09f2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e50f271-31", "ovs_interfaceid": "5e50f271-31a9-4755-a612-6b971179167a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1119.780615] env[62692]: DEBUG oslo_concurrency.lockutils [req-7afe8964-f151-4396-a4bf-f3bd26c33f06 req-72f778c6-a36c-4332-b98a-ac53e0ac32d6 service nova] Acquired lock "refresh_cache-e55a1a5b-3656-48d7-a610-04610b43999a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1119.780849] env[62692]: DEBUG nova.network.neutron [req-7afe8964-f151-4396-a4bf-f3bd26c33f06 req-72f778c6-a36c-4332-b98a-ac53e0ac32d6 service nova] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Refreshing network info cache for port 5e50f271-31a9-4755-a612-6b971179167a {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1119.782502] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:75:3c:39', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55764410-260e-4339-a020-6b30995584bf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5e50f271-31a9-4755-a612-6b971179167a', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1119.791352] env[62692]: DEBUG oslo.service.loopingcall [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1119.792716] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1119.793445] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.268s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.794073] env[62692]: DEBUG nova.compute.manager [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1119.797473] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-baf928d8-7357-450e-ab8b-57e12d364d14 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.814514] env[62692]: DEBUG oslo_concurrency.lockutils [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.196s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.815289] env[62692]: INFO nova.compute.claims [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1119.824913] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1119.824913] env[62692]: value = "task-1141860" [ 1119.824913] env[62692]: _type = "Task" [ 1119.824913] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.834804] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141860, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.214691] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141859, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.322424] env[62692]: DEBUG nova.compute.utils [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1120.324164] env[62692]: DEBUG nova.compute.manager [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1120.324379] env[62692]: DEBUG nova.network.neutron [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1120.338461] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141860, 'name': CreateVM_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.370731] env[62692]: DEBUG nova.policy [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '66364879aaea4bd2a4880b720da868ed', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '96b921004f8a4bf68640c3810fe7b819', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1120.623903] env[62692]: DEBUG nova.network.neutron [req-7afe8964-f151-4396-a4bf-f3bd26c33f06 req-72f778c6-a36c-4332-b98a-ac53e0ac32d6 service nova] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Updated VIF entry in instance network info cache for port 5e50f271-31a9-4755-a612-6b971179167a. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1120.624427] env[62692]: DEBUG nova.network.neutron [req-7afe8964-f151-4396-a4bf-f3bd26c33f06 req-72f778c6-a36c-4332-b98a-ac53e0ac32d6 service nova] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Updating instance_info_cache with network_info: [{"id": "5e50f271-31a9-4755-a612-6b971179167a", "address": "fa:16:3e:75:3c:39", "network": {"id": "dd260020-2d74-46ba-aa78-af29bdefa176", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-617179655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "589895ae20394e2fa525707bf7d09f2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e50f271-31", "ovs_interfaceid": "5e50f271-31a9-4755-a612-6b971179167a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1120.718406] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141859, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.599853} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.718682] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] dd8a0d53-5dc8-4b66-add7-f244e600468d/dd8a0d53-5dc8-4b66-add7-f244e600468d.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1120.718900] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1120.719196] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c86a1221-7c61-4a78-9ddc-5e6fe76bf27d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.727871] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1120.727871] env[62692]: value = "task-1141861" [ 1120.727871] env[62692]: _type = "Task" [ 1120.727871] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.740877] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141861, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.811168] env[62692]: DEBUG nova.network.neutron [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Successfully created port: fa8d79b4-2bf8-4b6b-a5c2-6893f618f420 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1120.830699] env[62692]: DEBUG nova.compute.manager [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1120.851739] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141860, 'name': CreateVM_Task, 'duration_secs': 0.813929} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.851973] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1120.852810] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1120.853010] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1120.853386] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1120.853720] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d49dcde9-c7e9-4afb-9c32-fa8ee4df2ec6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.864017] env[62692]: DEBUG oslo_vmware.api [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1120.864017] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52bf49fe-7332-7003-0838-4d80671747ce" [ 1120.864017] env[62692]: _type = "Task" [ 1120.864017] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.878008] env[62692]: DEBUG oslo_vmware.api [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52bf49fe-7332-7003-0838-4d80671747ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.022884] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25fbbfa9-58be-4095-90e7-ca8b8c605219 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.032747] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25731282-c657-4698-806b-edc270f272ce {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.069865] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7972f964-484c-4fd9-aace-c8d2dd338ddb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.079067] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22102398-cc56-4580-b062-011e20c08d05 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.094191] env[62692]: DEBUG nova.compute.provider_tree [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1121.128500] env[62692]: DEBUG oslo_concurrency.lockutils [req-7afe8964-f151-4396-a4bf-f3bd26c33f06 req-72f778c6-a36c-4332-b98a-ac53e0ac32d6 service nova] Releasing lock "refresh_cache-e55a1a5b-3656-48d7-a610-04610b43999a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1121.239277] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141861, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.098514} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.239616] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1121.240769] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad6f0096-da43-446a-b302-b15f5b316dae {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.267980] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] dd8a0d53-5dc8-4b66-add7-f244e600468d/dd8a0d53-5dc8-4b66-add7-f244e600468d.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1121.268345] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1454a50e-0af1-4120-9f46-1d79397f4349 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.290078] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1121.290078] env[62692]: value = "task-1141863" [ 1121.290078] env[62692]: _type = "Task" [ 1121.290078] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.301530] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141863, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.374985] env[62692]: DEBUG oslo_vmware.api [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52bf49fe-7332-7003-0838-4d80671747ce, 'name': SearchDatastore_Task, 'duration_secs': 0.018254} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.375333] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1121.375602] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1121.375815] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1121.375966] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.376171] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1121.376458] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-243a381a-25bb-4e05-9dfe-3f4968a95cf4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.386186] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1121.386439] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1121.388027] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-972901db-71d9-49fb-ab05-be24e5857790 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.395141] env[62692]: DEBUG oslo_vmware.api [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1121.395141] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52949683-7d81-5111-33e1-1d246fd738b3" [ 1121.395141] env[62692]: _type = "Task" [ 1121.395141] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.406851] env[62692]: DEBUG oslo_vmware.api [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52949683-7d81-5111-33e1-1d246fd738b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.598889] env[62692]: DEBUG nova.scheduler.client.report [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1121.802165] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141863, 'name': ReconfigVM_Task, 'duration_secs': 0.35612} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.802165] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Reconfigured VM instance instance-00000068 to attach disk [datastore1] dd8a0d53-5dc8-4b66-add7-f244e600468d/dd8a0d53-5dc8-4b66-add7-f244e600468d.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1121.802165] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4c9d915a-3898-4084-9f87-65b6cd7126ca {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.810835] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1121.810835] env[62692]: value = "task-1141864" [ 1121.810835] env[62692]: _type = "Task" [ 1121.810835] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.820765] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141864, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.848544] env[62692]: DEBUG nova.compute.manager [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1121.876153] env[62692]: DEBUG nova.virt.hardware [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1121.876426] env[62692]: DEBUG nova.virt.hardware [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1121.876592] env[62692]: DEBUG nova.virt.hardware [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1121.876782] env[62692]: DEBUG nova.virt.hardware [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1121.876933] env[62692]: DEBUG nova.virt.hardware [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1121.877108] env[62692]: DEBUG nova.virt.hardware [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1121.877331] env[62692]: DEBUG nova.virt.hardware [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1121.877494] env[62692]: DEBUG nova.virt.hardware [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1121.877664] env[62692]: DEBUG nova.virt.hardware [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1121.877827] env[62692]: DEBUG nova.virt.hardware [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1121.878000] env[62692]: DEBUG nova.virt.hardware [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1121.879279] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2da909f-9d21-4fe3-bc96-663fa60c66e2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.888289] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-061e8707-6dd7-4a0f-a0f7-25e4302b9aed {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.911912] env[62692]: DEBUG oslo_vmware.api [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52949683-7d81-5111-33e1-1d246fd738b3, 'name': SearchDatastore_Task, 'duration_secs': 0.011986} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.912781] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b375dcf-31e8-457a-952c-c9090531660d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.919141] env[62692]: DEBUG oslo_vmware.api [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1121.919141] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]523e055e-7c0c-cac7-aeea-b306326a1666" [ 1121.919141] env[62692]: _type = "Task" [ 1121.919141] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.928359] env[62692]: DEBUG oslo_vmware.api [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]523e055e-7c0c-cac7-aeea-b306326a1666, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.961393] env[62692]: DEBUG oslo_concurrency.lockutils [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "interface-94a07505-661f-4595-9024-8a9da9f2c2fc-613360e6-9cd1-4423-9c92-293618e0c0d1" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.961788] env[62692]: DEBUG oslo_concurrency.lockutils [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "interface-94a07505-661f-4595-9024-8a9da9f2c2fc-613360e6-9cd1-4423-9c92-293618e0c0d1" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.962191] env[62692]: DEBUG nova.objects.instance [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lazy-loading 'flavor' on Instance uuid 94a07505-661f-4595-9024-8a9da9f2c2fc {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1122.106088] env[62692]: DEBUG oslo_concurrency.lockutils [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.292s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.106719] env[62692]: DEBUG nova.compute.manager [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1122.322022] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141864, 'name': Rename_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.412305] env[62692]: DEBUG nova.compute.manager [req-639044a6-054d-4bbc-ab93-cdd4c18922f9 req-96b6cae6-d006-49f8-ae1e-8dfbdbd09846 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Received event network-vif-plugged-fa8d79b4-2bf8-4b6b-a5c2-6893f618f420 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1122.412558] env[62692]: DEBUG oslo_concurrency.lockutils [req-639044a6-054d-4bbc-ab93-cdd4c18922f9 req-96b6cae6-d006-49f8-ae1e-8dfbdbd09846 service nova] Acquiring lock "c2f2a485-363e-41af-9b3f-37092947ca08-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.412773] env[62692]: DEBUG oslo_concurrency.lockutils [req-639044a6-054d-4bbc-ab93-cdd4c18922f9 req-96b6cae6-d006-49f8-ae1e-8dfbdbd09846 service nova] Lock "c2f2a485-363e-41af-9b3f-37092947ca08-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.412948] env[62692]: DEBUG oslo_concurrency.lockutils [req-639044a6-054d-4bbc-ab93-cdd4c18922f9 req-96b6cae6-d006-49f8-ae1e-8dfbdbd09846 service nova] Lock "c2f2a485-363e-41af-9b3f-37092947ca08-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.413139] env[62692]: DEBUG nova.compute.manager [req-639044a6-054d-4bbc-ab93-cdd4c18922f9 req-96b6cae6-d006-49f8-ae1e-8dfbdbd09846 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] No waiting events found dispatching network-vif-plugged-fa8d79b4-2bf8-4b6b-a5c2-6893f618f420 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1122.413311] env[62692]: WARNING nova.compute.manager [req-639044a6-054d-4bbc-ab93-cdd4c18922f9 req-96b6cae6-d006-49f8-ae1e-8dfbdbd09846 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Received unexpected event network-vif-plugged-fa8d79b4-2bf8-4b6b-a5c2-6893f618f420 for instance with vm_state building and task_state spawning. [ 1122.430507] env[62692]: DEBUG oslo_vmware.api [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]523e055e-7c0c-cac7-aeea-b306326a1666, 'name': SearchDatastore_Task, 'duration_secs': 0.013333} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.430832] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1122.431019] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] e55a1a5b-3656-48d7-a610-04610b43999a/e55a1a5b-3656-48d7-a610-04610b43999a.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1122.431311] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b7eff6de-bcc3-4625-9b86-6fd237f07914 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.440488] env[62692]: DEBUG oslo_vmware.api [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1122.440488] env[62692]: value = "task-1141865" [ 1122.440488] env[62692]: _type = "Task" [ 1122.440488] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.449105] env[62692]: DEBUG oslo_vmware.api [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141865, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.589040] env[62692]: DEBUG nova.objects.instance [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lazy-loading 'pci_requests' on Instance uuid 94a07505-661f-4595-9024-8a9da9f2c2fc {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1122.610247] env[62692]: DEBUG nova.network.neutron [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Successfully updated port: fa8d79b4-2bf8-4b6b-a5c2-6893f618f420 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1122.612677] env[62692]: DEBUG nova.compute.utils [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1122.614156] env[62692]: DEBUG nova.compute.manager [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1122.614214] env[62692]: DEBUG nova.network.neutron [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1122.698444] env[62692]: DEBUG nova.policy [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '016134ac36fd4bd893226ae95a955e28', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37921f67664544478cb2568def068b11', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1122.822397] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141864, 'name': Rename_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.953839] env[62692]: DEBUG oslo_vmware.api [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141865, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.093832] env[62692]: DEBUG nova.objects.base [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Object Instance<94a07505-661f-4595-9024-8a9da9f2c2fc> lazy-loaded attributes: flavor,pci_requests {{(pid=62692) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1123.094061] env[62692]: DEBUG nova.network.neutron [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1123.115334] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "refresh_cache-c2f2a485-363e-41af-9b3f-37092947ca08" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1123.115508] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquired lock "refresh_cache-c2f2a485-363e-41af-9b3f-37092947ca08" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1123.115669] env[62692]: DEBUG nova.network.neutron [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1123.117371] env[62692]: DEBUG nova.compute.manager [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1123.210223] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-32cc40cc-426b-46bb-9325-d895f694d4fc tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Volume attach. Driver type: vmdk {{(pid=62692) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1123.210223] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-32cc40cc-426b-46bb-9325-d895f694d4fc tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249058', 'volume_id': '02a5babd-e299-40a6-a55f-e70273241e04', 'name': 'volume-02a5babd-e299-40a6-a55f-e70273241e04', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970', 'attached_at': '', 'detached_at': '', 'volume_id': '02a5babd-e299-40a6-a55f-e70273241e04', 'serial': '02a5babd-e299-40a6-a55f-e70273241e04'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1123.210223] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dcfbd29-cf09-4ce9-a633-e47e2122646a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.215191] env[62692]: DEBUG nova.policy [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a251ef5f6cb143178928babc75a0fef0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ba97a775e99a463c845518b0df8cd3e2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1123.235667] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-793509e5-c67f-411f-9ad6-ca302a122e7c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.239633] env[62692]: DEBUG nova.network.neutron [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Successfully created port: 5c943de3-e4d0-479b-9d1d-b38816b891b2 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1123.267695] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-32cc40cc-426b-46bb-9325-d895f694d4fc tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] volume-02a5babd-e299-40a6-a55f-e70273241e04/volume-02a5babd-e299-40a6-a55f-e70273241e04.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1123.268323] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9865b247-a47f-4583-b0a4-02f04e4b298f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.289908] env[62692]: DEBUG oslo_vmware.api [None req-32cc40cc-426b-46bb-9325-d895f694d4fc tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1123.289908] env[62692]: value = "task-1141866" [ 1123.289908] env[62692]: _type = "Task" [ 1123.289908] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.301517] env[62692]: DEBUG oslo_vmware.api [None req-32cc40cc-426b-46bb-9325-d895f694d4fc tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141866, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.321960] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141864, 'name': Rename_Task, 'duration_secs': 1.205476} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.325026] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1123.325026] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-90a85f32-d3bf-4bc0-8355-54af27f535fd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.335025] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1123.335025] env[62692]: value = "task-1141867" [ 1123.335025] env[62692]: _type = "Task" [ 1123.335025] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.348366] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141867, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.457916] env[62692]: DEBUG oslo_vmware.api [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141865, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.001995} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.458436] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] e55a1a5b-3656-48d7-a610-04610b43999a/e55a1a5b-3656-48d7-a610-04610b43999a.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1123.458825] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1123.459208] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c43891eb-6d7a-4401-ae85-e51517cd4fe7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.468975] env[62692]: DEBUG oslo_vmware.api [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1123.468975] env[62692]: value = "task-1141868" [ 1123.468975] env[62692]: _type = "Task" [ 1123.468975] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.479321] env[62692]: DEBUG oslo_vmware.api [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141868, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.690690] env[62692]: DEBUG nova.network.neutron [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1123.801817] env[62692]: DEBUG oslo_vmware.api [None req-32cc40cc-426b-46bb-9325-d895f694d4fc tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141866, 'name': ReconfigVM_Task, 'duration_secs': 0.436593} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.802341] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-32cc40cc-426b-46bb-9325-d895f694d4fc tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Reconfigured VM instance instance-00000060 to attach disk [datastore1] volume-02a5babd-e299-40a6-a55f-e70273241e04/volume-02a5babd-e299-40a6-a55f-e70273241e04.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1123.810643] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c9d091e-53f2-4753-8268-3a37b65b268d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.828991] env[62692]: DEBUG oslo_vmware.api [None req-32cc40cc-426b-46bb-9325-d895f694d4fc tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1123.828991] env[62692]: value = "task-1141869" [ 1123.828991] env[62692]: _type = "Task" [ 1123.828991] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.841812] env[62692]: DEBUG oslo_vmware.api [None req-32cc40cc-426b-46bb-9325-d895f694d4fc tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141869, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.848745] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141867, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.946249] env[62692]: DEBUG nova.network.neutron [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Updating instance_info_cache with network_info: [{"id": "fa8d79b4-2bf8-4b6b-a5c2-6893f618f420", "address": "fa:16:3e:a0:b7:ee", "network": {"id": "22795f7f-18d7-4b10-8051-bcff4e0af3ac", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-445901728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96b921004f8a4bf68640c3810fe7b819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa8d79b4-2b", "ovs_interfaceid": "fa8d79b4-2bf8-4b6b-a5c2-6893f618f420", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1123.980359] env[62692]: DEBUG oslo_vmware.api [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141868, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.134191} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.980664] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1123.981519] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97922c50-4a61-4466-b986-3b2fb437a7e9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.009341] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] e55a1a5b-3656-48d7-a610-04610b43999a/e55a1a5b-3656-48d7-a610-04610b43999a.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1124.010112] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd354267-ce8a-463d-b6b7-83ad76539096 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.033217] env[62692]: DEBUG oslo_vmware.api [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1124.033217] env[62692]: value = "task-1141870" [ 1124.033217] env[62692]: _type = "Task" [ 1124.033217] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.044247] env[62692]: DEBUG oslo_vmware.api [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141870, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.134256] env[62692]: DEBUG nova.compute.manager [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1124.157605] env[62692]: DEBUG nova.virt.hardware [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1124.157900] env[62692]: DEBUG nova.virt.hardware [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1124.158079] env[62692]: DEBUG nova.virt.hardware [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1124.158280] env[62692]: DEBUG nova.virt.hardware [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1124.158495] env[62692]: DEBUG nova.virt.hardware [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1124.158604] env[62692]: DEBUG nova.virt.hardware [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1124.159069] env[62692]: DEBUG nova.virt.hardware [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1124.159069] env[62692]: DEBUG nova.virt.hardware [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1124.159229] env[62692]: DEBUG nova.virt.hardware [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1124.159386] env[62692]: DEBUG nova.virt.hardware [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1124.159522] env[62692]: DEBUG nova.virt.hardware [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1124.160916] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8268ee4f-a883-4adc-b4f7-a7fc25dbe9e9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.170069] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b49b418-0472-4ada-9e32-7a3f9dd9d51d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.347327] env[62692]: DEBUG oslo_vmware.api [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141867, 'name': PowerOnVM_Task, 'duration_secs': 0.730578} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.350354] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1124.350579] env[62692]: INFO nova.compute.manager [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Took 9.58 seconds to spawn the instance on the hypervisor. [ 1124.350769] env[62692]: DEBUG nova.compute.manager [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1124.351074] env[62692]: DEBUG oslo_vmware.api [None req-32cc40cc-426b-46bb-9325-d895f694d4fc tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141869, 'name': ReconfigVM_Task, 'duration_secs': 0.149453} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.351809] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d809c778-dc15-4048-98e2-31829c4018d4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.355109] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-32cc40cc-426b-46bb-9325-d895f694d4fc tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249058', 'volume_id': '02a5babd-e299-40a6-a55f-e70273241e04', 'name': 'volume-02a5babd-e299-40a6-a55f-e70273241e04', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970', 'attached_at': '', 'detached_at': '', 'volume_id': '02a5babd-e299-40a6-a55f-e70273241e04', 'serial': '02a5babd-e299-40a6-a55f-e70273241e04'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1124.449828] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Releasing lock "refresh_cache-c2f2a485-363e-41af-9b3f-37092947ca08" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1124.450179] env[62692]: DEBUG nova.compute.manager [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Instance network_info: |[{"id": "fa8d79b4-2bf8-4b6b-a5c2-6893f618f420", "address": "fa:16:3e:a0:b7:ee", "network": {"id": "22795f7f-18d7-4b10-8051-bcff4e0af3ac", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-445901728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96b921004f8a4bf68640c3810fe7b819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa8d79b4-2b", "ovs_interfaceid": "fa8d79b4-2bf8-4b6b-a5c2-6893f618f420", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1124.450623] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a0:b7:ee', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd1da5fc2-0280-4f76-ac97-20ea4bc7bb16', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fa8d79b4-2bf8-4b6b-a5c2-6893f618f420', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1124.458008] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Creating folder: Project (96b921004f8a4bf68640c3810fe7b819). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1124.458322] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2ee85254-20b5-479a-a16d-354e48a8d69b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.470470] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Created folder: Project (96b921004f8a4bf68640c3810fe7b819) in parent group-v248868. [ 1124.470650] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Creating folder: Instances. Parent ref: group-v249060. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1124.470852] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-188eb192-7283-4d9e-929e-3ec609ed21c9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.482762] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Created folder: Instances in parent group-v249060. [ 1124.482762] env[62692]: DEBUG oslo.service.loopingcall [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1124.482762] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1124.482762] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3575c67d-4868-4a5d-9940-bdcf22bee6a4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.500060] env[62692]: DEBUG nova.compute.manager [req-38cc6b92-057f-45b4-9c2f-b9780b7c107d req-58c89e21-ce7a-4529-8bff-f95578b5321a service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Received event network-changed-fa8d79b4-2bf8-4b6b-a5c2-6893f618f420 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1124.500255] env[62692]: DEBUG nova.compute.manager [req-38cc6b92-057f-45b4-9c2f-b9780b7c107d req-58c89e21-ce7a-4529-8bff-f95578b5321a service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Refreshing instance network info cache due to event network-changed-fa8d79b4-2bf8-4b6b-a5c2-6893f618f420. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1124.500484] env[62692]: DEBUG oslo_concurrency.lockutils [req-38cc6b92-057f-45b4-9c2f-b9780b7c107d req-58c89e21-ce7a-4529-8bff-f95578b5321a service nova] Acquiring lock "refresh_cache-c2f2a485-363e-41af-9b3f-37092947ca08" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.500626] env[62692]: DEBUG oslo_concurrency.lockutils [req-38cc6b92-057f-45b4-9c2f-b9780b7c107d req-58c89e21-ce7a-4529-8bff-f95578b5321a service nova] Acquired lock "refresh_cache-c2f2a485-363e-41af-9b3f-37092947ca08" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.500843] env[62692]: DEBUG nova.network.neutron [req-38cc6b92-057f-45b4-9c2f-b9780b7c107d req-58c89e21-ce7a-4529-8bff-f95578b5321a service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Refreshing network info cache for port fa8d79b4-2bf8-4b6b-a5c2-6893f618f420 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1124.510366] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1124.510366] env[62692]: value = "task-1141873" [ 1124.510366] env[62692]: _type = "Task" [ 1124.510366] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.519353] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141873, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.545220] env[62692]: DEBUG oslo_vmware.api [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141870, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.879091] env[62692]: INFO nova.compute.manager [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Took 16.05 seconds to build instance. [ 1125.019693] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141873, 'name': CreateVM_Task, 'duration_secs': 0.379338} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.020015] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1125.021186] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.021186] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.021798] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1125.022593] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e42e101-c48c-4288-9f23-01730d543e29 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.028581] env[62692]: DEBUG oslo_vmware.api [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1125.028581] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52815f55-6d3f-25de-975c-94bd8e781f0e" [ 1125.028581] env[62692]: _type = "Task" [ 1125.028581] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.036427] env[62692]: DEBUG nova.network.neutron [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Successfully updated port: 613360e6-9cd1-4423-9c92-293618e0c0d1 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1125.047815] env[62692]: DEBUG oslo_vmware.api [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52815f55-6d3f-25de-975c-94bd8e781f0e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.050605] env[62692]: DEBUG oslo_vmware.api [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141870, 'name': ReconfigVM_Task, 'duration_secs': 0.957355} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.051572] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Reconfigured VM instance instance-00000069 to attach disk [datastore1] e55a1a5b-3656-48d7-a610-04610b43999a/e55a1a5b-3656-48d7-a610-04610b43999a.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1125.052231] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bce342da-4e0f-4cba-b4f0-b1516632481d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.062189] env[62692]: DEBUG oslo_vmware.api [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1125.062189] env[62692]: value = "task-1141874" [ 1125.062189] env[62692]: _type = "Task" [ 1125.062189] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.074547] env[62692]: DEBUG oslo_vmware.api [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141874, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.085754] env[62692]: DEBUG nova.network.neutron [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Successfully updated port: 5c943de3-e4d0-479b-9d1d-b38816b891b2 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1125.242364] env[62692]: DEBUG nova.network.neutron [req-38cc6b92-057f-45b4-9c2f-b9780b7c107d req-58c89e21-ce7a-4529-8bff-f95578b5321a service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Updated VIF entry in instance network info cache for port fa8d79b4-2bf8-4b6b-a5c2-6893f618f420. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1125.242364] env[62692]: DEBUG nova.network.neutron [req-38cc6b92-057f-45b4-9c2f-b9780b7c107d req-58c89e21-ce7a-4529-8bff-f95578b5321a service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Updating instance_info_cache with network_info: [{"id": "fa8d79b4-2bf8-4b6b-a5c2-6893f618f420", "address": "fa:16:3e:a0:b7:ee", "network": {"id": "22795f7f-18d7-4b10-8051-bcff4e0af3ac", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-445901728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96b921004f8a4bf68640c3810fe7b819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa8d79b4-2b", "ovs_interfaceid": "fa8d79b4-2bf8-4b6b-a5c2-6893f618f420", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1125.265253] env[62692]: DEBUG nova.compute.manager [req-2d67b864-af59-4442-80c5-5a7d1e03cc18 req-53b5d00e-6364-4681-85a8-8cf74a9c5248 service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Received event network-vif-plugged-613360e6-9cd1-4423-9c92-293618e0c0d1 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1125.265499] env[62692]: DEBUG oslo_concurrency.lockutils [req-2d67b864-af59-4442-80c5-5a7d1e03cc18 req-53b5d00e-6364-4681-85a8-8cf74a9c5248 service nova] Acquiring lock "94a07505-661f-4595-9024-8a9da9f2c2fc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.265720] env[62692]: DEBUG oslo_concurrency.lockutils [req-2d67b864-af59-4442-80c5-5a7d1e03cc18 req-53b5d00e-6364-4681-85a8-8cf74a9c5248 service nova] Lock "94a07505-661f-4595-9024-8a9da9f2c2fc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.265918] env[62692]: DEBUG oslo_concurrency.lockutils [req-2d67b864-af59-4442-80c5-5a7d1e03cc18 req-53b5d00e-6364-4681-85a8-8cf74a9c5248 service nova] Lock "94a07505-661f-4595-9024-8a9da9f2c2fc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.266428] env[62692]: DEBUG nova.compute.manager [req-2d67b864-af59-4442-80c5-5a7d1e03cc18 req-53b5d00e-6364-4681-85a8-8cf74a9c5248 service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] No waiting events found dispatching network-vif-plugged-613360e6-9cd1-4423-9c92-293618e0c0d1 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1125.266710] env[62692]: WARNING nova.compute.manager [req-2d67b864-af59-4442-80c5-5a7d1e03cc18 req-53b5d00e-6364-4681-85a8-8cf74a9c5248 service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Received unexpected event network-vif-plugged-613360e6-9cd1-4423-9c92-293618e0c0d1 for instance with vm_state active and task_state None. [ 1125.382831] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a8a4cc50-8180-4f77-9f38-f3dd34652c6f tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Lock "dd8a0d53-5dc8-4b66-add7-f244e600468d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.561s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.400422] env[62692]: DEBUG nova.objects.instance [None req-32cc40cc-426b-46bb-9325-d895f694d4fc tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lazy-loading 'flavor' on Instance uuid a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1125.542800] env[62692]: DEBUG oslo_concurrency.lockutils [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "refresh_cache-94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.542996] env[62692]: DEBUG oslo_concurrency.lockutils [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "refresh_cache-94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.543102] env[62692]: DEBUG nova.network.neutron [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1125.544558] env[62692]: DEBUG oslo_vmware.api [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52815f55-6d3f-25de-975c-94bd8e781f0e, 'name': SearchDatastore_Task, 'duration_secs': 0.020517} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.545025] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.545107] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1125.545904] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.546070] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.546513] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1125.547172] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8bfbdb69-9636-446f-b648-79802c0afd0d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.560481] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1125.561048] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1125.561861] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-961a8d02-e61c-419a-9535-a0c81c3f3a2d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.579151] env[62692]: DEBUG oslo_vmware.api [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141874, 'name': Rename_Task, 'duration_secs': 0.47053} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.579830] env[62692]: DEBUG oslo_vmware.api [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1125.579830] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52980384-5383-47bb-682e-734efb50bf7f" [ 1125.579830] env[62692]: _type = "Task" [ 1125.579830] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.580093] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1125.580417] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-46a693ed-1fad-443c-85a5-7ea1b764b60b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.595905] env[62692]: DEBUG oslo_concurrency.lockutils [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "refresh_cache-1c7ce408-6e9c-43bd-8d81-6acb12353ed8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.596126] env[62692]: DEBUG oslo_concurrency.lockutils [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquired lock "refresh_cache-1c7ce408-6e9c-43bd-8d81-6acb12353ed8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.596292] env[62692]: DEBUG nova.network.neutron [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1125.597543] env[62692]: DEBUG oslo_vmware.api [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52980384-5383-47bb-682e-734efb50bf7f, 'name': SearchDatastore_Task, 'duration_secs': 0.012778} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.600272] env[62692]: DEBUG oslo_vmware.api [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1125.600272] env[62692]: value = "task-1141875" [ 1125.600272] env[62692]: _type = "Task" [ 1125.600272] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.600457] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5442849f-69cc-4c3c-8cbc-ad84a52bc0dc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.614588] env[62692]: DEBUG oslo_vmware.api [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141875, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.614950] env[62692]: DEBUG oslo_vmware.api [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1125.614950] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52469939-bc11-3e7f-109c-b412aff3c73d" [ 1125.614950] env[62692]: _type = "Task" [ 1125.614950] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.625721] env[62692]: DEBUG oslo_vmware.api [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52469939-bc11-3e7f-109c-b412aff3c73d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.649620] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Acquiring lock "c20139ae-5076-4f6b-b4f0-047332559acf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.649871] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Lock "c20139ae-5076-4f6b-b4f0-047332559acf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.745574] env[62692]: DEBUG oslo_concurrency.lockutils [req-38cc6b92-057f-45b4-9c2f-b9780b7c107d req-58c89e21-ce7a-4529-8bff-f95578b5321a service nova] Releasing lock "refresh_cache-c2f2a485-363e-41af-9b3f-37092947ca08" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.906035] env[62692]: DEBUG oslo_concurrency.lockutils [None req-32cc40cc-426b-46bb-9325-d895f694d4fc tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.329s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.083644] env[62692]: WARNING nova.network.neutron [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] b34af32c-59b7-4de4-85ef-08bd67c1e2a2 already exists in list: networks containing: ['b34af32c-59b7-4de4-85ef-08bd67c1e2a2']. ignoring it [ 1126.115544] env[62692]: DEBUG oslo_vmware.api [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141875, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.130018] env[62692]: DEBUG oslo_vmware.api [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52469939-bc11-3e7f-109c-b412aff3c73d, 'name': SearchDatastore_Task, 'duration_secs': 0.015417} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.130018] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1126.130018] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] c2f2a485-363e-41af-9b3f-37092947ca08/c2f2a485-363e-41af-9b3f-37092947ca08.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1126.130018] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d2f1e5b5-3d4b-46cd-9395-ea1623fbc7c8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.139581] env[62692]: DEBUG oslo_vmware.api [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1126.139581] env[62692]: value = "task-1141876" [ 1126.139581] env[62692]: _type = "Task" [ 1126.139581] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.144043] env[62692]: DEBUG nova.network.neutron [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1126.153745] env[62692]: DEBUG nova.compute.manager [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1126.156334] env[62692]: DEBUG oslo_vmware.api [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1141876, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.403638] env[62692]: DEBUG nova.network.neutron [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Updating instance_info_cache with network_info: [{"id": "5c943de3-e4d0-479b-9d1d-b38816b891b2", "address": "fa:16:3e:54:d8:95", "network": {"id": "2febb660-329d-403c-8c2c-f11c0f5f547c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1806650357-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37921f67664544478cb2568def068b11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c943de3-e4", "ovs_interfaceid": "5c943de3-e4d0-479b-9d1d-b38816b891b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1126.471728] env[62692]: DEBUG nova.compute.manager [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Stashing vm_state: active {{(pid=62692) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1126.600367] env[62692]: DEBUG nova.network.neutron [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Updating instance_info_cache with network_info: [{"id": "d51ffd3f-c023-45e2-8bfa-1424b7ccb188", "address": "fa:16:3e:a8:44:f1", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd51ffd3f-c0", "ovs_interfaceid": "d51ffd3f-c023-45e2-8bfa-1424b7ccb188", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "613360e6-9cd1-4423-9c92-293618e0c0d1", "address": "fa:16:3e:4e:4b:47", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap613360e6-9c", "ovs_interfaceid": "613360e6-9cd1-4423-9c92-293618e0c0d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1126.616317] env[62692]: DEBUG oslo_vmware.api [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141875, 'name': PowerOnVM_Task, 'duration_secs': 0.81764} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.617113] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1126.617113] env[62692]: INFO nova.compute.manager [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Took 9.32 seconds to spawn the instance on the hypervisor. [ 1126.617113] env[62692]: DEBUG nova.compute.manager [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1126.617857] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3816d93-2931-40d3-9313-55373bf81754 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.652035] env[62692]: DEBUG oslo_vmware.api [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1141876, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.684173] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1126.684173] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.684846] env[62692]: INFO nova.compute.claims [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1126.812430] env[62692]: INFO nova.compute.manager [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Rescuing [ 1126.812741] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquiring lock "refresh_cache-dd8a0d53-5dc8-4b66-add7-f244e600468d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1126.812919] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquired lock "refresh_cache-dd8a0d53-5dc8-4b66-add7-f244e600468d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.813117] env[62692]: DEBUG nova.network.neutron [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1126.906724] env[62692]: DEBUG oslo_concurrency.lockutils [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Releasing lock "refresh_cache-1c7ce408-6e9c-43bd-8d81-6acb12353ed8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1126.907144] env[62692]: DEBUG nova.compute.manager [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Instance network_info: |[{"id": "5c943de3-e4d0-479b-9d1d-b38816b891b2", "address": "fa:16:3e:54:d8:95", "network": {"id": "2febb660-329d-403c-8c2c-f11c0f5f547c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1806650357-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37921f67664544478cb2568def068b11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c943de3-e4", "ovs_interfaceid": "5c943de3-e4d0-479b-9d1d-b38816b891b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1126.907658] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:54:d8:95', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5c943de3-e4d0-479b-9d1d-b38816b891b2', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1126.916502] env[62692]: DEBUG oslo.service.loopingcall [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1126.916749] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1126.916977] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-424b65ce-01ae-4fa3-8018-d51fc193a56f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.933351] env[62692]: DEBUG oslo_concurrency.lockutils [None req-38c3f90a-5920-43e8-8610-816c6ace64af tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1126.933618] env[62692]: DEBUG oslo_concurrency.lockutils [None req-38c3f90a-5920-43e8-8610-816c6ace64af tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.938055] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1126.938055] env[62692]: value = "task-1141877" [ 1126.938055] env[62692]: _type = "Task" [ 1126.938055] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.947493] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141877, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.998118] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.001493] env[62692]: DEBUG nova.compute.manager [req-9fc67b2c-8791-4e02-b1f7-a346b084ce2f req-6e89ec55-b353-4e34-8553-23252fb2c449 service nova] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Received event network-vif-plugged-5c943de3-e4d0-479b-9d1d-b38816b891b2 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1127.001810] env[62692]: DEBUG oslo_concurrency.lockutils [req-9fc67b2c-8791-4e02-b1f7-a346b084ce2f req-6e89ec55-b353-4e34-8553-23252fb2c449 service nova] Acquiring lock "1c7ce408-6e9c-43bd-8d81-6acb12353ed8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.002137] env[62692]: DEBUG oslo_concurrency.lockutils [req-9fc67b2c-8791-4e02-b1f7-a346b084ce2f req-6e89ec55-b353-4e34-8553-23252fb2c449 service nova] Lock "1c7ce408-6e9c-43bd-8d81-6acb12353ed8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.002337] env[62692]: DEBUG oslo_concurrency.lockutils [req-9fc67b2c-8791-4e02-b1f7-a346b084ce2f req-6e89ec55-b353-4e34-8553-23252fb2c449 service nova] Lock "1c7ce408-6e9c-43bd-8d81-6acb12353ed8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.002553] env[62692]: DEBUG nova.compute.manager [req-9fc67b2c-8791-4e02-b1f7-a346b084ce2f req-6e89ec55-b353-4e34-8553-23252fb2c449 service nova] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] No waiting events found dispatching network-vif-plugged-5c943de3-e4d0-479b-9d1d-b38816b891b2 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1127.002736] env[62692]: WARNING nova.compute.manager [req-9fc67b2c-8791-4e02-b1f7-a346b084ce2f req-6e89ec55-b353-4e34-8553-23252fb2c449 service nova] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Received unexpected event network-vif-plugged-5c943de3-e4d0-479b-9d1d-b38816b891b2 for instance with vm_state building and task_state spawning. [ 1127.002899] env[62692]: DEBUG nova.compute.manager [req-9fc67b2c-8791-4e02-b1f7-a346b084ce2f req-6e89ec55-b353-4e34-8553-23252fb2c449 service nova] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Received event network-changed-5c943de3-e4d0-479b-9d1d-b38816b891b2 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1127.003127] env[62692]: DEBUG nova.compute.manager [req-9fc67b2c-8791-4e02-b1f7-a346b084ce2f req-6e89ec55-b353-4e34-8553-23252fb2c449 service nova] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Refreshing instance network info cache due to event network-changed-5c943de3-e4d0-479b-9d1d-b38816b891b2. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1127.003391] env[62692]: DEBUG oslo_concurrency.lockutils [req-9fc67b2c-8791-4e02-b1f7-a346b084ce2f req-6e89ec55-b353-4e34-8553-23252fb2c449 service nova] Acquiring lock "refresh_cache-1c7ce408-6e9c-43bd-8d81-6acb12353ed8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1127.003563] env[62692]: DEBUG oslo_concurrency.lockutils [req-9fc67b2c-8791-4e02-b1f7-a346b084ce2f req-6e89ec55-b353-4e34-8553-23252fb2c449 service nova] Acquired lock "refresh_cache-1c7ce408-6e9c-43bd-8d81-6acb12353ed8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.003761] env[62692]: DEBUG nova.network.neutron [req-9fc67b2c-8791-4e02-b1f7-a346b084ce2f req-6e89ec55-b353-4e34-8553-23252fb2c449 service nova] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Refreshing network info cache for port 5c943de3-e4d0-479b-9d1d-b38816b891b2 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1127.103080] env[62692]: DEBUG oslo_concurrency.lockutils [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "refresh_cache-94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1127.104044] env[62692]: DEBUG oslo_concurrency.lockutils [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1127.104236] env[62692]: DEBUG oslo_concurrency.lockutils [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.105167] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fd69491-52d1-4f87-b20b-e513a55021f2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.124837] env[62692]: DEBUG nova.virt.hardware [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1127.125138] env[62692]: DEBUG nova.virt.hardware [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1127.125307] env[62692]: DEBUG nova.virt.hardware [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1127.125497] env[62692]: DEBUG nova.virt.hardware [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1127.125653] env[62692]: DEBUG nova.virt.hardware [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1127.125802] env[62692]: DEBUG nova.virt.hardware [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1127.126032] env[62692]: DEBUG nova.virt.hardware [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1127.126201] env[62692]: DEBUG nova.virt.hardware [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1127.126377] env[62692]: DEBUG nova.virt.hardware [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1127.126548] env[62692]: DEBUG nova.virt.hardware [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1127.126726] env[62692]: DEBUG nova.virt.hardware [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1127.133406] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Reconfiguring VM to attach interface {{(pid=62692) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1127.137590] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd081d74-49bf-417c-ae2a-950110f37eb1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.152642] env[62692]: INFO nova.compute.manager [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Took 17.18 seconds to build instance. [ 1127.163244] env[62692]: DEBUG oslo_vmware.api [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1141876, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.675661} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.164849] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] c2f2a485-363e-41af-9b3f-37092947ca08/c2f2a485-363e-41af-9b3f-37092947ca08.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1127.165113] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1127.165454] env[62692]: DEBUG oslo_vmware.api [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1127.165454] env[62692]: value = "task-1141878" [ 1127.165454] env[62692]: _type = "Task" [ 1127.165454] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.165721] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1080d93c-73f6-4b5d-a7ed-564f73fea2d2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.177490] env[62692]: DEBUG oslo_vmware.api [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141878, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.178796] env[62692]: DEBUG oslo_vmware.api [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1127.178796] env[62692]: value = "task-1141879" [ 1127.178796] env[62692]: _type = "Task" [ 1127.178796] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.193692] env[62692]: DEBUG oslo_vmware.api [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1141879, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.365608] env[62692]: DEBUG nova.compute.manager [req-40028aeb-23a4-43f7-a249-8eac2118bde0 req-51dd7033-1f8d-49a2-a2da-a5bd31cdc4d8 service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Received event network-changed-613360e6-9cd1-4423-9c92-293618e0c0d1 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1127.365608] env[62692]: DEBUG nova.compute.manager [req-40028aeb-23a4-43f7-a249-8eac2118bde0 req-51dd7033-1f8d-49a2-a2da-a5bd31cdc4d8 service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Refreshing instance network info cache due to event network-changed-613360e6-9cd1-4423-9c92-293618e0c0d1. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1127.365608] env[62692]: DEBUG oslo_concurrency.lockutils [req-40028aeb-23a4-43f7-a249-8eac2118bde0 req-51dd7033-1f8d-49a2-a2da-a5bd31cdc4d8 service nova] Acquiring lock "refresh_cache-94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1127.365608] env[62692]: DEBUG oslo_concurrency.lockutils [req-40028aeb-23a4-43f7-a249-8eac2118bde0 req-51dd7033-1f8d-49a2-a2da-a5bd31cdc4d8 service nova] Acquired lock "refresh_cache-94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.365928] env[62692]: DEBUG nova.network.neutron [req-40028aeb-23a4-43f7-a249-8eac2118bde0 req-51dd7033-1f8d-49a2-a2da-a5bd31cdc4d8 service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Refreshing network info cache for port 613360e6-9cd1-4423-9c92-293618e0c0d1 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1127.436894] env[62692]: INFO nova.compute.manager [None req-38c3f90a-5920-43e8-8610-816c6ace64af tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Detaching volume 02a5babd-e299-40a6-a55f-e70273241e04 [ 1127.451849] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141877, 'name': CreateVM_Task, 'duration_secs': 0.390795} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.454922] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1127.455819] env[62692]: DEBUG oslo_concurrency.lockutils [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1127.456107] env[62692]: DEBUG oslo_concurrency.lockutils [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.456490] env[62692]: DEBUG oslo_concurrency.lockutils [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1127.457110] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf745d4e-94c7-4c4c-be9a-cbefab12a538 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.463282] env[62692]: DEBUG oslo_vmware.api [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1127.463282] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]521656dd-d886-4ec9-6c29-088dcaf892d7" [ 1127.463282] env[62692]: _type = "Task" [ 1127.463282] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.473490] env[62692]: DEBUG oslo_vmware.api [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]521656dd-d886-4ec9-6c29-088dcaf892d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.476805] env[62692]: INFO nova.virt.block_device [None req-38c3f90a-5920-43e8-8610-816c6ace64af tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Attempting to driver detach volume 02a5babd-e299-40a6-a55f-e70273241e04 from mountpoint /dev/sdb [ 1127.477053] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-38c3f90a-5920-43e8-8610-816c6ace64af tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Volume detach. Driver type: vmdk {{(pid=62692) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1127.477258] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-38c3f90a-5920-43e8-8610-816c6ace64af tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249058', 'volume_id': '02a5babd-e299-40a6-a55f-e70273241e04', 'name': 'volume-02a5babd-e299-40a6-a55f-e70273241e04', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970', 'attached_at': '', 'detached_at': '', 'volume_id': '02a5babd-e299-40a6-a55f-e70273241e04', 'serial': '02a5babd-e299-40a6-a55f-e70273241e04'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1127.478072] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7006ac99-84b2-4987-963e-9e98ee1c3f40 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.504499] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2d252bc-2ec1-409b-b8c1-37cc8b18f4e1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.517496] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3db8fb8b-000d-4012-9139-5b5fb68151e2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.540359] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb741ccf-1791-403f-8c82-509b48b3687d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.557341] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-38c3f90a-5920-43e8-8610-816c6ace64af tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] The volume has not been displaced from its original location: [datastore1] volume-02a5babd-e299-40a6-a55f-e70273241e04/volume-02a5babd-e299-40a6-a55f-e70273241e04.vmdk. No consolidation needed. {{(pid=62692) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1127.563070] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-38c3f90a-5920-43e8-8610-816c6ace64af tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Reconfiguring VM instance instance-00000060 to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1127.566062] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2df5d048-a0c9-48f3-b916-99306457e694 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.580769] env[62692]: DEBUG nova.network.neutron [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Updating instance_info_cache with network_info: [{"id": "f8a6a1d2-a31b-451a-854c-8efc6e7aa879", "address": "fa:16:3e:9f:bb:e7", "network": {"id": "50c1330e-1a5f-42d1-9987-79ddcc0ad7aa", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-813901268-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b1c44eb72b334ee3add29026d312e37b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f92f0b92-d6fb-4d00-8ad5-6b3809ed5493", "external-id": "nsx-vlan-transportzone-48", "segmentation_id": 48, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8a6a1d2-a3", "ovs_interfaceid": "f8a6a1d2-a31b-451a-854c-8efc6e7aa879", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1127.588767] env[62692]: DEBUG oslo_vmware.api [None req-38c3f90a-5920-43e8-8610-816c6ace64af tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1127.588767] env[62692]: value = "task-1141880" [ 1127.588767] env[62692]: _type = "Task" [ 1127.588767] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.599031] env[62692]: DEBUG oslo_vmware.api [None req-38c3f90a-5920-43e8-8610-816c6ace64af tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141880, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.658716] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ab67c886-790a-4da5-8b23-03c8e1cd9176 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "e55a1a5b-3656-48d7-a610-04610b43999a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.694s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.680829] env[62692]: DEBUG oslo_vmware.api [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141878, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.690523] env[62692]: DEBUG oslo_vmware.api [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1141879, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.132711} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.690596] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1127.691543] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e73ab384-5fb1-4554-b576-6a4f6c187952 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.723573] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] c2f2a485-363e-41af-9b3f-37092947ca08/c2f2a485-363e-41af-9b3f-37092947ca08.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1127.726918] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90266194-ebdc-4e03-8c7c-b9706dfa781c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.753056] env[62692]: DEBUG oslo_vmware.api [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1127.753056] env[62692]: value = "task-1141881" [ 1127.753056] env[62692]: _type = "Task" [ 1127.753056] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.764827] env[62692]: DEBUG oslo_vmware.api [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1141881, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.787861] env[62692]: DEBUG nova.network.neutron [req-9fc67b2c-8791-4e02-b1f7-a346b084ce2f req-6e89ec55-b353-4e34-8553-23252fb2c449 service nova] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Updated VIF entry in instance network info cache for port 5c943de3-e4d0-479b-9d1d-b38816b891b2. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1127.788361] env[62692]: DEBUG nova.network.neutron [req-9fc67b2c-8791-4e02-b1f7-a346b084ce2f req-6e89ec55-b353-4e34-8553-23252fb2c449 service nova] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Updating instance_info_cache with network_info: [{"id": "5c943de3-e4d0-479b-9d1d-b38816b891b2", "address": "fa:16:3e:54:d8:95", "network": {"id": "2febb660-329d-403c-8c2c-f11c0f5f547c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1806650357-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37921f67664544478cb2568def068b11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c943de3-e4", "ovs_interfaceid": "5c943de3-e4d0-479b-9d1d-b38816b891b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1127.924793] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d4c86d2-1924-408f-bb21-e1f182104f4b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.934624] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a46b490-2949-47d3-b21b-f437183f76b9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.978347] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e9ceabd-1cbb-4fe6-8d87-54d32952223b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.992716] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2b92714-956b-4543-acf4-943762729f76 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.998179] env[62692]: DEBUG oslo_vmware.api [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]521656dd-d886-4ec9-6c29-088dcaf892d7, 'name': SearchDatastore_Task, 'duration_secs': 0.015952} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.998805] env[62692]: DEBUG oslo_concurrency.lockutils [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1127.998805] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1127.999121] env[62692]: DEBUG oslo_concurrency.lockutils [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1127.999376] env[62692]: DEBUG oslo_concurrency.lockutils [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.999656] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1128.003086] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-119ff214-3bfb-49a4-b239-5c7188404f18 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.014386] env[62692]: DEBUG nova.compute.provider_tree [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1128.025210] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1128.025514] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1128.027336] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17fd22a1-3b68-43e1-82aa-fe5747176d81 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.035203] env[62692]: DEBUG oslo_vmware.api [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1128.035203] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52370337-1bc1-e58b-dadf-9e6282831f95" [ 1128.035203] env[62692]: _type = "Task" [ 1128.035203] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.049136] env[62692]: DEBUG oslo_vmware.api [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52370337-1bc1-e58b-dadf-9e6282831f95, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.083758] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Releasing lock "refresh_cache-dd8a0d53-5dc8-4b66-add7-f244e600468d" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1128.102444] env[62692]: DEBUG oslo_vmware.api [None req-38c3f90a-5920-43e8-8610-816c6ace64af tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141880, 'name': ReconfigVM_Task, 'duration_secs': 0.275693} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.104611] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-38c3f90a-5920-43e8-8610-816c6ace64af tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Reconfigured VM instance instance-00000060 to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1128.109691] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bca2b773-6ed7-4e73-95d5-a5cf205ca81a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.134521] env[62692]: DEBUG oslo_vmware.api [None req-38c3f90a-5920-43e8-8610-816c6ace64af tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1128.134521] env[62692]: value = "task-1141882" [ 1128.134521] env[62692]: _type = "Task" [ 1128.134521] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.144240] env[62692]: DEBUG oslo_vmware.api [None req-38c3f90a-5920-43e8-8610-816c6ace64af tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141882, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.166148] env[62692]: DEBUG nova.network.neutron [req-40028aeb-23a4-43f7-a249-8eac2118bde0 req-51dd7033-1f8d-49a2-a2da-a5bd31cdc4d8 service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Updated VIF entry in instance network info cache for port 613360e6-9cd1-4423-9c92-293618e0c0d1. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1128.166724] env[62692]: DEBUG nova.network.neutron [req-40028aeb-23a4-43f7-a249-8eac2118bde0 req-51dd7033-1f8d-49a2-a2da-a5bd31cdc4d8 service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Updating instance_info_cache with network_info: [{"id": "d51ffd3f-c023-45e2-8bfa-1424b7ccb188", "address": "fa:16:3e:a8:44:f1", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd51ffd3f-c0", "ovs_interfaceid": "d51ffd3f-c023-45e2-8bfa-1424b7ccb188", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "613360e6-9cd1-4423-9c92-293618e0c0d1", "address": "fa:16:3e:4e:4b:47", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap613360e6-9c", "ovs_interfaceid": "613360e6-9cd1-4423-9c92-293618e0c0d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1128.178078] env[62692]: DEBUG oslo_vmware.api [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141878, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.249317] env[62692]: DEBUG oslo_vmware.rw_handles [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ea7a3a-3def-d36e-02ce-55b9740d4384/disk-0.vmdk. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1128.250562] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50b6ec74-36cf-48bd-b76d-2ba1d5f9e690 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.263227] env[62692]: DEBUG oslo_vmware.rw_handles [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ea7a3a-3def-d36e-02ce-55b9740d4384/disk-0.vmdk is in state: ready. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1128.263447] env[62692]: ERROR oslo_vmware.rw_handles [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ea7a3a-3def-d36e-02ce-55b9740d4384/disk-0.vmdk due to incomplete transfer. [ 1128.268014] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-16148d78-155c-431f-b5f7-673999d8aff7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.268233] env[62692]: DEBUG oslo_vmware.api [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1141881, 'name': ReconfigVM_Task, 'duration_secs': 0.331126} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.268489] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Reconfigured VM instance instance-0000006a to attach disk [datastore1] c2f2a485-363e-41af-9b3f-37092947ca08/c2f2a485-363e-41af-9b3f-37092947ca08.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1128.269513] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c662e467-8316-4ec0-b13f-f32d8f7ad815 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.278710] env[62692]: DEBUG oslo_vmware.api [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1128.278710] env[62692]: value = "task-1141883" [ 1128.278710] env[62692]: _type = "Task" [ 1128.278710] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.288316] env[62692]: DEBUG oslo_vmware.api [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1141883, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.289574] env[62692]: DEBUG oslo_vmware.rw_handles [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ea7a3a-3def-d36e-02ce-55b9740d4384/disk-0.vmdk. {{(pid=62692) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1128.289834] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Uploaded image b7b0878a-b2c2-4474-9dbb-47b954aa8a86 to the Glance image server {{(pid=62692) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1128.292197] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Destroying the VM {{(pid=62692) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1128.292703] env[62692]: DEBUG oslo_concurrency.lockutils [req-9fc67b2c-8791-4e02-b1f7-a346b084ce2f req-6e89ec55-b353-4e34-8553-23252fb2c449 service nova] Releasing lock "refresh_cache-1c7ce408-6e9c-43bd-8d81-6acb12353ed8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1128.293111] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f41db8b6-b629-47fa-aeca-ed60a9a8d61d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.301487] env[62692]: DEBUG oslo_vmware.api [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1128.301487] env[62692]: value = "task-1141884" [ 1128.301487] env[62692]: _type = "Task" [ 1128.301487] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.312419] env[62692]: DEBUG oslo_vmware.api [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141884, 'name': Destroy_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.518114] env[62692]: DEBUG nova.scheduler.client.report [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1128.545238] env[62692]: DEBUG oslo_vmware.api [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52370337-1bc1-e58b-dadf-9e6282831f95, 'name': SearchDatastore_Task, 'duration_secs': 0.022055} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.547230] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4fae741d-aa11-45bb-a610-a49819630d80 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.558527] env[62692]: DEBUG oslo_vmware.api [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1128.558527] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5295cca4-3b4c-73b8-193e-8e86c7541791" [ 1128.558527] env[62692]: _type = "Task" [ 1128.558527] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.567186] env[62692]: DEBUG oslo_vmware.api [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5295cca4-3b4c-73b8-193e-8e86c7541791, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.628585] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1128.628846] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-824c89f0-4e5f-41df-b434-aed10e07932f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.636821] env[62692]: DEBUG oslo_vmware.api [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1128.636821] env[62692]: value = "task-1141885" [ 1128.636821] env[62692]: _type = "Task" [ 1128.636821] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.651626] env[62692]: DEBUG oslo_vmware.api [None req-38c3f90a-5920-43e8-8610-816c6ace64af tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141882, 'name': ReconfigVM_Task, 'duration_secs': 0.163375} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.651920] env[62692]: DEBUG oslo_vmware.api [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141885, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.652282] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-38c3f90a-5920-43e8-8610-816c6ace64af tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249058', 'volume_id': '02a5babd-e299-40a6-a55f-e70273241e04', 'name': 'volume-02a5babd-e299-40a6-a55f-e70273241e04', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970', 'attached_at': '', 'detached_at': '', 'volume_id': '02a5babd-e299-40a6-a55f-e70273241e04', 'serial': '02a5babd-e299-40a6-a55f-e70273241e04'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1128.672119] env[62692]: DEBUG oslo_concurrency.lockutils [req-40028aeb-23a4-43f7-a249-8eac2118bde0 req-51dd7033-1f8d-49a2-a2da-a5bd31cdc4d8 service nova] Releasing lock "refresh_cache-94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1128.677933] env[62692]: DEBUG oslo_vmware.api [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141878, 'name': ReconfigVM_Task, 'duration_secs': 1.127427} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.678524] env[62692]: DEBUG oslo_concurrency.lockutils [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1128.678654] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Reconfigured VM to attach interface {{(pid=62692) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1128.790275] env[62692]: DEBUG oslo_vmware.api [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1141883, 'name': Rename_Task, 'duration_secs': 0.178521} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.790640] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1128.790950] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dbff90ca-bf06-4b1b-a324-511fb1419288 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.807287] env[62692]: DEBUG oslo_vmware.api [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1128.807287] env[62692]: value = "task-1141886" [ 1128.807287] env[62692]: _type = "Task" [ 1128.807287] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.813302] env[62692]: DEBUG oslo_vmware.api [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141884, 'name': Destroy_Task} progress is 33%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.818768] env[62692]: DEBUG oslo_vmware.api [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1141886, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.023526] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.340s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.023877] env[62692]: DEBUG nova.compute.manager [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1129.026617] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 2.029s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.064553] env[62692]: DEBUG nova.compute.manager [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Stashing vm_state: active {{(pid=62692) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1129.076077] env[62692]: DEBUG oslo_vmware.api [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5295cca4-3b4c-73b8-193e-8e86c7541791, 'name': SearchDatastore_Task, 'duration_secs': 0.031576} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.076342] env[62692]: DEBUG oslo_concurrency.lockutils [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1129.076600] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 1c7ce408-6e9c-43bd-8d81-6acb12353ed8/1c7ce408-6e9c-43bd-8d81-6acb12353ed8.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1129.076864] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0edb1a13-afd8-4f13-a8a7-e6ab6b1787da {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.084733] env[62692]: DEBUG oslo_vmware.api [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1129.084733] env[62692]: value = "task-1141887" [ 1129.084733] env[62692]: _type = "Task" [ 1129.084733] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.093672] env[62692]: DEBUG oslo_vmware.api [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141887, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.150746] env[62692]: DEBUG oslo_vmware.api [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141885, 'name': PowerOffVM_Task, 'duration_secs': 0.383585} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.151122] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1129.152096] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26ed0ffe-ab7c-4468-89e3-9e0130f6567e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.173742] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d027c335-a3d8-4466-ac01-782e533c5bc3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.182714] env[62692]: DEBUG oslo_concurrency.lockutils [None req-db9199e1-0a11-4111-ba4c-3781589b872c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "interface-94a07505-661f-4595-9024-8a9da9f2c2fc-613360e6-9cd1-4423-9c92-293618e0c0d1" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.221s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.206961] env[62692]: DEBUG nova.objects.instance [None req-38c3f90a-5920-43e8-8610-816c6ace64af tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lazy-loading 'flavor' on Instance uuid a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1129.211534] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1129.211534] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba3f94fb-f13c-481a-86cf-4c8b5d247415 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.219083] env[62692]: DEBUG oslo_vmware.api [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1129.219083] env[62692]: value = "task-1141888" [ 1129.219083] env[62692]: _type = "Task" [ 1129.219083] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.229919] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] VM already powered off {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1129.230159] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1129.230436] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1129.230596] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.230783] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1129.231027] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5a193eb8-d259-4e62-9863-0fca0d2c86d8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.240089] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1129.240296] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1129.241057] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d57a59cf-d4a8-4d1e-993a-d3deee1fbfd5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.247940] env[62692]: DEBUG oslo_vmware.api [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1129.247940] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52285032-1d43-6319-960d-abe1bd1bf79d" [ 1129.247940] env[62692]: _type = "Task" [ 1129.247940] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.256615] env[62692]: DEBUG oslo_vmware.api [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52285032-1d43-6319-960d-abe1bd1bf79d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.315185] env[62692]: DEBUG oslo_vmware.api [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141884, 'name': Destroy_Task, 'duration_secs': 0.795961} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.315915] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Destroyed the VM [ 1129.316269] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Deleting Snapshot of the VM instance {{(pid=62692) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1129.316542] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b7124982-0631-4c7b-bb82-988f1c576c00 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.322016] env[62692]: DEBUG oslo_vmware.api [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1141886, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.329018] env[62692]: DEBUG oslo_vmware.api [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1129.329018] env[62692]: value = "task-1141889" [ 1129.329018] env[62692]: _type = "Task" [ 1129.329018] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.338193] env[62692]: DEBUG oslo_vmware.api [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141889, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.530455] env[62692]: DEBUG nova.compute.utils [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1129.535108] env[62692]: INFO nova.compute.claims [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1129.538904] env[62692]: DEBUG nova.compute.manager [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1129.539162] env[62692]: DEBUG nova.network.neutron [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1129.583967] env[62692]: DEBUG nova.policy [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0c7418503ce543c0a08521a3f9715c44', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a331bc93fe1140b68242feba6b02655a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1129.586894] env[62692]: DEBUG oslo_concurrency.lockutils [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.599884] env[62692]: DEBUG oslo_vmware.api [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141887, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.759007] env[62692]: DEBUG oslo_vmware.api [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52285032-1d43-6319-960d-abe1bd1bf79d, 'name': SearchDatastore_Task, 'duration_secs': 0.023499} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.760075] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc6ccd7b-371a-42b6-862b-0626601b3c1d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.766815] env[62692]: DEBUG oslo_vmware.api [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1129.766815] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]527c8d14-7f7b-8a73-5557-177a4140cc87" [ 1129.766815] env[62692]: _type = "Task" [ 1129.766815] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.776017] env[62692]: DEBUG oslo_vmware.api [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]527c8d14-7f7b-8a73-5557-177a4140cc87, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.818800] env[62692]: DEBUG oslo_vmware.api [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1141886, 'name': PowerOnVM_Task, 'duration_secs': 0.634824} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.819097] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1129.819311] env[62692]: INFO nova.compute.manager [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Took 7.97 seconds to spawn the instance on the hypervisor. [ 1129.819494] env[62692]: DEBUG nova.compute.manager [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1129.820414] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bddafd7-2736-42c2-8e09-98b8cd788c56 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.837197] env[62692]: DEBUG oslo_vmware.api [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141889, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.885332] env[62692]: DEBUG nova.network.neutron [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Successfully created port: e93a1d6d-aa66-47ac-8d7c-db90c1d090f8 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1130.039940] env[62692]: DEBUG nova.compute.manager [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1130.044472] env[62692]: INFO nova.compute.resource_tracker [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Updating resource usage from migration 7933eebe-a34a-46f0-8c3c-80776a5c4adc [ 1130.098120] env[62692]: DEBUG oslo_vmware.api [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141887, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.687821} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.098414] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 1c7ce408-6e9c-43bd-8d81-6acb12353ed8/1c7ce408-6e9c-43bd-8d81-6acb12353ed8.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1130.098669] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1130.098929] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-addcd3fd-7302-4cb9-8255-e208826d5d1e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.109737] env[62692]: DEBUG oslo_vmware.api [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1130.109737] env[62692]: value = "task-1141890" [ 1130.109737] env[62692]: _type = "Task" [ 1130.109737] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.121617] env[62692]: DEBUG oslo_vmware.api [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141890, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.218452] env[62692]: DEBUG oslo_concurrency.lockutils [None req-38c3f90a-5920-43e8-8610-816c6ace64af tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.285s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.268188] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29d87d95-7598-40d2-9b79-3bf9cb0272d7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.283106] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9de1ea7-b23d-429b-9c0f-09922826d0e1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.286418] env[62692]: DEBUG oslo_vmware.api [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]527c8d14-7f7b-8a73-5557-177a4140cc87, 'name': SearchDatastore_Task, 'duration_secs': 0.011577} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.286690] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1130.286987] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] dd8a0d53-5dc8-4b66-add7-f244e600468d/81eb64fb-5a0c-47b7-9948-7dcb01b911bd-rescue.vmdk. {{(pid=62692) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1130.287591] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a1434edf-c045-49c5-bf26-32ee4b1727e9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.321331] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cd5bb2c-5517-46d3-95b9-35f4e9760a91 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.324089] env[62692]: DEBUG oslo_vmware.api [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1130.324089] env[62692]: value = "task-1141891" [ 1130.324089] env[62692]: _type = "Task" [ 1130.324089] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.338249] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37dfd335-09c3-4a1a-9382-1dd51956a8fe {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.346647] env[62692]: DEBUG oslo_vmware.api [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141891, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.347104] env[62692]: INFO nova.compute.manager [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Took 15.82 seconds to build instance. [ 1130.353659] env[62692]: DEBUG oslo_vmware.api [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141889, 'name': RemoveSnapshot_Task} progress is 36%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.362392] env[62692]: DEBUG nova.compute.provider_tree [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1130.623628] env[62692]: DEBUG oslo_vmware.api [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141890, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075311} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.623628] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1130.624517] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-367fbd2d-9d2f-4149-bc1c-0edc4ae1cd85 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.648050] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] 1c7ce408-6e9c-43bd-8d81-6acb12353ed8/1c7ce408-6e9c-43bd-8d81-6acb12353ed8.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1130.648315] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cc9692c8-aeb4-4c9f-acc4-a8027a8cbb4a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.669689] env[62692]: DEBUG oslo_vmware.api [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1130.669689] env[62692]: value = "task-1141892" [ 1130.669689] env[62692]: _type = "Task" [ 1130.669689] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.682506] env[62692]: DEBUG oslo_vmware.api [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141892, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.836246] env[62692]: DEBUG oslo_vmware.api [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141891, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.846069] env[62692]: DEBUG oslo_vmware.api [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141889, 'name': RemoveSnapshot_Task, 'duration_secs': 1.507956} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.846534] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Deleted Snapshot of the VM instance {{(pid=62692) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1130.846903] env[62692]: INFO nova.compute.manager [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Took 17.85 seconds to snapshot the instance on the hypervisor. [ 1130.850623] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2202748a-a444-4a91-aae9-92c8e46fca74 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "c2f2a485-363e-41af-9b3f-37092947ca08" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.339s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.865692] env[62692]: DEBUG nova.scheduler.client.report [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1130.920541] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "interface-94a07505-661f-4595-9024-8a9da9f2c2fc-613360e6-9cd1-4423-9c92-293618e0c0d1" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.920807] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "interface-94a07505-661f-4595-9024-8a9da9f2c2fc-613360e6-9cd1-4423-9c92-293618e0c0d1" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.054684] env[62692]: DEBUG nova.compute.manager [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1131.081306] env[62692]: DEBUG nova.virt.hardware [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1131.081643] env[62692]: DEBUG nova.virt.hardware [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1131.081813] env[62692]: DEBUG nova.virt.hardware [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1131.082015] env[62692]: DEBUG nova.virt.hardware [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1131.082268] env[62692]: DEBUG nova.virt.hardware [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1131.082614] env[62692]: DEBUG nova.virt.hardware [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1131.082932] env[62692]: DEBUG nova.virt.hardware [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1131.083128] env[62692]: DEBUG nova.virt.hardware [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1131.083322] env[62692]: DEBUG nova.virt.hardware [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1131.083502] env[62692]: DEBUG nova.virt.hardware [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1131.083698] env[62692]: DEBUG nova.virt.hardware [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1131.084655] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b67c7c2e-641c-42b2-acf9-c2f8c89030a5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.093840] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55b4bc5a-ad42-4d5f-81a5-c3ea6c121fb4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.180535] env[62692]: DEBUG oslo_vmware.api [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141892, 'name': ReconfigVM_Task, 'duration_secs': 0.304219} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.180784] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Reconfigured VM instance instance-0000006b to attach disk [datastore2] 1c7ce408-6e9c-43bd-8d81-6acb12353ed8/1c7ce408-6e9c-43bd-8d81-6acb12353ed8.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1131.181527] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-01881459-140c-4f3d-9de7-be86c34d7424 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.189887] env[62692]: DEBUG oslo_vmware.api [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1131.189887] env[62692]: value = "task-1141893" [ 1131.189887] env[62692]: _type = "Task" [ 1131.189887] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.200763] env[62692]: DEBUG oslo_vmware.api [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141893, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.249887] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.250373] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.250731] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.251218] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.251501] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.257040] env[62692]: DEBUG nova.compute.manager [req-2729a611-e337-440b-8a2f-f250aa8917d0 req-bdec9792-193d-40b8-b94b-2206d3836c49 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Received event network-changed-fa8d79b4-2bf8-4b6b-a5c2-6893f618f420 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1131.257040] env[62692]: DEBUG nova.compute.manager [req-2729a611-e337-440b-8a2f-f250aa8917d0 req-bdec9792-193d-40b8-b94b-2206d3836c49 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Refreshing instance network info cache due to event network-changed-fa8d79b4-2bf8-4b6b-a5c2-6893f618f420. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1131.257159] env[62692]: DEBUG oslo_concurrency.lockutils [req-2729a611-e337-440b-8a2f-f250aa8917d0 req-bdec9792-193d-40b8-b94b-2206d3836c49 service nova] Acquiring lock "refresh_cache-c2f2a485-363e-41af-9b3f-37092947ca08" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.257306] env[62692]: DEBUG oslo_concurrency.lockutils [req-2729a611-e337-440b-8a2f-f250aa8917d0 req-bdec9792-193d-40b8-b94b-2206d3836c49 service nova] Acquired lock "refresh_cache-c2f2a485-363e-41af-9b3f-37092947ca08" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.257738] env[62692]: DEBUG nova.network.neutron [req-2729a611-e337-440b-8a2f-f250aa8917d0 req-bdec9792-193d-40b8-b94b-2206d3836c49 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Refreshing network info cache for port fa8d79b4-2bf8-4b6b-a5c2-6893f618f420 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1131.259267] env[62692]: INFO nova.compute.manager [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Terminating instance [ 1131.262145] env[62692]: DEBUG nova.compute.manager [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1131.262349] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1131.263597] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5a79198-740f-49a7-834a-89fbe49602cb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.274534] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1131.275207] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-23c221c4-3a11-41aa-96b3-e6efaf405fe5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.282669] env[62692]: DEBUG oslo_vmware.api [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1131.282669] env[62692]: value = "task-1141894" [ 1131.282669] env[62692]: _type = "Task" [ 1131.282669] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.293908] env[62692]: DEBUG oslo_vmware.api [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141894, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.336507] env[62692]: DEBUG oslo_vmware.api [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141891, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.996851} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.336870] env[62692]: INFO nova.virt.vmwareapi.ds_util [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] dd8a0d53-5dc8-4b66-add7-f244e600468d/81eb64fb-5a0c-47b7-9948-7dcb01b911bd-rescue.vmdk. [ 1131.337698] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-079b652b-dfcd-47d3-be4a-957f615b930e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.371246] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] dd8a0d53-5dc8-4b66-add7-f244e600468d/81eb64fb-5a0c-47b7-9948-7dcb01b911bd-rescue.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1131.373946] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82f541be-7f55-44bb-96a2-1fbb47b87245 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.391763] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.365s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.392134] env[62692]: INFO nova.compute.manager [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Migrating [ 1131.399601] env[62692]: DEBUG oslo_concurrency.lockutils [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.813s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.414819] env[62692]: DEBUG oslo_vmware.api [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1131.414819] env[62692]: value = "task-1141895" [ 1131.414819] env[62692]: _type = "Task" [ 1131.414819] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.424637] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.424894] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.426652] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d951afb3-0ad5-4e3c-bc70-f815269860a4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.434835] env[62692]: DEBUG nova.compute.manager [None req-ccf32652-7d66-4dcb-9aa7-6449a0a08af7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Found 1 images (rotation: 2) {{(pid=62692) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1131.437760] env[62692]: DEBUG oslo_vmware.api [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141895, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.454713] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fc016a6-1b39-4c11-bd7d-2c19300c70d8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.484052] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Reconfiguring VM to detach interface {{(pid=62692) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1131.484534] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-764adfc3-7324-49bf-bdb9-bf56bef7204a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.509062] env[62692]: DEBUG oslo_vmware.api [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1131.509062] env[62692]: value = "task-1141896" [ 1131.509062] env[62692]: _type = "Task" [ 1131.509062] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.519649] env[62692]: DEBUG oslo_vmware.api [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141896, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.702276] env[62692]: DEBUG oslo_vmware.api [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141893, 'name': Rename_Task, 'duration_secs': 0.207511} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.702276] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1131.702610] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aa26bba6-dbf2-427d-a83c-b32e1d650ae2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.715046] env[62692]: DEBUG oslo_vmware.api [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1131.715046] env[62692]: value = "task-1141897" [ 1131.715046] env[62692]: _type = "Task" [ 1131.715046] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.725993] env[62692]: DEBUG oslo_vmware.api [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141897, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.780739] env[62692]: DEBUG nova.compute.manager [req-7496cf0c-c303-4bae-af9b-baded72e5131 req-8c97183e-1f83-4812-99eb-d986ca5368f7 service nova] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Received event network-vif-plugged-e93a1d6d-aa66-47ac-8d7c-db90c1d090f8 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1131.780739] env[62692]: DEBUG oslo_concurrency.lockutils [req-7496cf0c-c303-4bae-af9b-baded72e5131 req-8c97183e-1f83-4812-99eb-d986ca5368f7 service nova] Acquiring lock "c20139ae-5076-4f6b-b4f0-047332559acf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.780739] env[62692]: DEBUG oslo_concurrency.lockutils [req-7496cf0c-c303-4bae-af9b-baded72e5131 req-8c97183e-1f83-4812-99eb-d986ca5368f7 service nova] Lock "c20139ae-5076-4f6b-b4f0-047332559acf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.780739] env[62692]: DEBUG oslo_concurrency.lockutils [req-7496cf0c-c303-4bae-af9b-baded72e5131 req-8c97183e-1f83-4812-99eb-d986ca5368f7 service nova] Lock "c20139ae-5076-4f6b-b4f0-047332559acf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.781473] env[62692]: DEBUG nova.compute.manager [req-7496cf0c-c303-4bae-af9b-baded72e5131 req-8c97183e-1f83-4812-99eb-d986ca5368f7 service nova] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] No waiting events found dispatching network-vif-plugged-e93a1d6d-aa66-47ac-8d7c-db90c1d090f8 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1131.781983] env[62692]: WARNING nova.compute.manager [req-7496cf0c-c303-4bae-af9b-baded72e5131 req-8c97183e-1f83-4812-99eb-d986ca5368f7 service nova] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Received unexpected event network-vif-plugged-e93a1d6d-aa66-47ac-8d7c-db90c1d090f8 for instance with vm_state building and task_state spawning. [ 1131.800021] env[62692]: DEBUG oslo_vmware.api [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141894, 'name': PowerOffVM_Task, 'duration_secs': 0.318097} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.800021] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1131.800021] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1131.800021] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d59107b2-2d6b-4003-aee2-4b545cf04efe {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.885852] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1131.886034] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1131.886235] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Deleting the datastore file [datastore2] a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1131.886505] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-47d540cd-ad6a-41e3-b8b7-a00be26898be {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.895073] env[62692]: DEBUG oslo_vmware.api [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1131.895073] env[62692]: value = "task-1141899" [ 1131.895073] env[62692]: _type = "Task" [ 1131.895073] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.909659] env[62692]: INFO nova.compute.claims [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1131.913842] env[62692]: DEBUG oslo_vmware.api [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141899, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.918105] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "refresh_cache-cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.918359] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired lock "refresh_cache-cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.922281] env[62692]: DEBUG nova.network.neutron [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1131.941329] env[62692]: DEBUG oslo_vmware.api [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141895, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.022518] env[62692]: DEBUG oslo_vmware.api [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141896, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.157420] env[62692]: DEBUG nova.network.neutron [req-2729a611-e337-440b-8a2f-f250aa8917d0 req-bdec9792-193d-40b8-b94b-2206d3836c49 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Updated VIF entry in instance network info cache for port fa8d79b4-2bf8-4b6b-a5c2-6893f618f420. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1132.157852] env[62692]: DEBUG nova.network.neutron [req-2729a611-e337-440b-8a2f-f250aa8917d0 req-bdec9792-193d-40b8-b94b-2206d3836c49 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Updating instance_info_cache with network_info: [{"id": "fa8d79b4-2bf8-4b6b-a5c2-6893f618f420", "address": "fa:16:3e:a0:b7:ee", "network": {"id": "22795f7f-18d7-4b10-8051-bcff4e0af3ac", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-445901728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.250", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96b921004f8a4bf68640c3810fe7b819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa8d79b4-2b", "ovs_interfaceid": "fa8d79b4-2bf8-4b6b-a5c2-6893f618f420", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1132.211191] env[62692]: DEBUG nova.compute.manager [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1132.212203] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a94a060-2ae7-4e86-8ae7-4ed1dbb9fef1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.229282] env[62692]: DEBUG oslo_vmware.api [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141897, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.321571] env[62692]: DEBUG nova.network.neutron [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Successfully updated port: e93a1d6d-aa66-47ac-8d7c-db90c1d090f8 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1132.407287] env[62692]: DEBUG oslo_vmware.api [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141899, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.29323} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.407639] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1132.407858] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1132.408059] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1132.408240] env[62692]: INFO nova.compute.manager [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1132.408620] env[62692]: DEBUG oslo.service.loopingcall [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1132.408910] env[62692]: DEBUG nova.compute.manager [-] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1132.409080] env[62692]: DEBUG nova.network.neutron [-] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1132.415805] env[62692]: INFO nova.compute.resource_tracker [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Updating resource usage from migration 34373c4c-3481-421d-b142-1919d2e93950 [ 1132.436021] env[62692]: DEBUG oslo_vmware.api [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141895, 'name': ReconfigVM_Task, 'duration_secs': 0.836643} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.436396] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Reconfigured VM instance instance-00000068 to attach disk [datastore1] dd8a0d53-5dc8-4b66-add7-f244e600468d/81eb64fb-5a0c-47b7-9948-7dcb01b911bd-rescue.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1132.437625] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8784cdb6-acf6-4f49-9f78-f2bd741829ea {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.471592] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af16283d-d0f5-490c-ba7f-0bb705cee925 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.489923] env[62692]: DEBUG oslo_vmware.api [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1132.489923] env[62692]: value = "task-1141900" [ 1132.489923] env[62692]: _type = "Task" [ 1132.489923] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.504473] env[62692]: DEBUG oslo_vmware.api [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141900, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.522972] env[62692]: DEBUG oslo_vmware.api [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141896, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.660826] env[62692]: DEBUG oslo_concurrency.lockutils [req-2729a611-e337-440b-8a2f-f250aa8917d0 req-bdec9792-193d-40b8-b94b-2206d3836c49 service nova] Releasing lock "refresh_cache-c2f2a485-363e-41af-9b3f-37092947ca08" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.692466] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81368126-3524-4341-86fa-61b5efc71ec3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.701521] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-658d304c-f391-4345-af92-64965b597fea {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.742414] env[62692]: INFO nova.compute.manager [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] instance snapshotting [ 1132.742414] env[62692]: DEBUG nova.objects.instance [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lazy-loading 'flavor' on Instance uuid 90932f05-fad3-4c6d-87ae-ab059351b0be {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1132.751278] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5f8af55-755b-4f13-8ca1-c0c6e806e3a3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.760961] env[62692]: DEBUG oslo_vmware.api [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141897, 'name': PowerOnVM_Task, 'duration_secs': 0.796702} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.763923] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1132.764191] env[62692]: INFO nova.compute.manager [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Took 8.63 seconds to spawn the instance on the hypervisor. [ 1132.764393] env[62692]: DEBUG nova.compute.manager [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1132.765525] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8e470f4-d05c-4032-b5fe-a8712895f42d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.768899] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-647aac88-27e3-47b8-8cf1-46342d0feec2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.793557] env[62692]: DEBUG nova.compute.provider_tree [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1132.825733] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Acquiring lock "refresh_cache-c20139ae-5076-4f6b-b4f0-047332559acf" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1132.825919] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Acquired lock "refresh_cache-c20139ae-5076-4f6b-b4f0-047332559acf" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1132.826106] env[62692]: DEBUG nova.network.neutron [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1132.961766] env[62692]: DEBUG nova.network.neutron [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Updating instance_info_cache with network_info: [{"id": "a3885c51-6f7b-41e8-8948-41fac442a288", "address": "fa:16:3e:e4:9b:6f", "network": {"id": "f70727d6-a9aa-44b0-a244-2533b3f96571", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-284105837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1037e5d436bd429391159b87a712f6e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3885c51-6f", "ovs_interfaceid": "a3885c51-6f7b-41e8-8948-41fac442a288", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1133.002853] env[62692]: DEBUG oslo_vmware.api [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141900, 'name': ReconfigVM_Task, 'duration_secs': 0.467582} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.003226] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1133.003382] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-69ed97ef-0770-4ca1-b82a-b22ad5a0b0c0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.013206] env[62692]: DEBUG oslo_vmware.api [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1133.013206] env[62692]: value = "task-1141901" [ 1133.013206] env[62692]: _type = "Task" [ 1133.013206] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.026322] env[62692]: DEBUG oslo_vmware.api [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141896, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.029234] env[62692]: DEBUG oslo_vmware.api [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141901, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.211894] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "7183ec4a-992d-48f1-8dda-7f499c2f4e1d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.212316] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "7183ec4a-992d-48f1-8dda-7f499c2f4e1d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.212371] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "7183ec4a-992d-48f1-8dda-7f499c2f4e1d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.212559] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "7183ec4a-992d-48f1-8dda-7f499c2f4e1d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.212756] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "7183ec4a-992d-48f1-8dda-7f499c2f4e1d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.218638] env[62692]: INFO nova.compute.manager [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Terminating instance [ 1133.224329] env[62692]: DEBUG nova.compute.manager [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1133.224514] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1133.225361] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68c7c646-92a6-46a5-9ec4-6f74b0365222 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.234524] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1133.234939] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5fb843da-8b0a-42c7-bd4d-168be468c694 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.243199] env[62692]: DEBUG oslo_vmware.api [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1133.243199] env[62692]: value = "task-1141902" [ 1133.243199] env[62692]: _type = "Task" [ 1133.243199] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.252280] env[62692]: DEBUG oslo_vmware.api [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141902, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.259114] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-354accc9-d08e-4c99-8138-54397f89400d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.282689] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75af55f9-7249-480c-a175-60cb363963a3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.300069] env[62692]: DEBUG nova.scheduler.client.report [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1133.307178] env[62692]: DEBUG nova.compute.manager [req-37ff6ccd-9e01-411f-b9d8-2db1b92184d8 req-c87f53cf-2cf1-4785-a021-d7664ba82f20 service nova] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Received event network-changed-e93a1d6d-aa66-47ac-8d7c-db90c1d090f8 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1133.307178] env[62692]: DEBUG nova.compute.manager [req-37ff6ccd-9e01-411f-b9d8-2db1b92184d8 req-c87f53cf-2cf1-4785-a021-d7664ba82f20 service nova] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Refreshing instance network info cache due to event network-changed-e93a1d6d-aa66-47ac-8d7c-db90c1d090f8. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1133.307178] env[62692]: DEBUG oslo_concurrency.lockutils [req-37ff6ccd-9e01-411f-b9d8-2db1b92184d8 req-c87f53cf-2cf1-4785-a021-d7664ba82f20 service nova] Acquiring lock "refresh_cache-c20139ae-5076-4f6b-b4f0-047332559acf" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1133.307796] env[62692]: INFO nova.compute.manager [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Took 18.70 seconds to build instance. [ 1133.390731] env[62692]: DEBUG nova.network.neutron [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1133.465474] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Releasing lock "refresh_cache-cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1133.525211] env[62692]: DEBUG oslo_vmware.api [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141896, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.531278] env[62692]: DEBUG oslo_vmware.api [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141901, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.638952] env[62692]: DEBUG nova.network.neutron [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Updating instance_info_cache with network_info: [{"id": "e93a1d6d-aa66-47ac-8d7c-db90c1d090f8", "address": "fa:16:3e:e7:6a:de", "network": {"id": "647be031-b44d-4439-a059-a40feedbbccf", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-477265227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a331bc93fe1140b68242feba6b02655a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f246b87-f105-4b33-a71d-5caf8e99e074", "external-id": "nsx-vlan-transportzone-583", "segmentation_id": 583, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape93a1d6d-aa", "ovs_interfaceid": "e93a1d6d-aa66-47ac-8d7c-db90c1d090f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1133.678393] env[62692]: DEBUG nova.network.neutron [-] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1133.753904] env[62692]: DEBUG oslo_vmware.api [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141902, 'name': PowerOffVM_Task, 'duration_secs': 0.293691} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.754462] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1133.754710] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1133.754987] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-155e7201-2c57-4536-b30b-af967fc5669a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.796089] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Creating Snapshot of the VM instance {{(pid=62692) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1133.796433] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-cc8a53b8-8723-4c7a-8569-b6d4e5c41a42 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.805648] env[62692]: DEBUG oslo_vmware.api [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1133.805648] env[62692]: value = "task-1141904" [ 1133.805648] env[62692]: _type = "Task" [ 1133.805648] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.809824] env[62692]: DEBUG oslo_concurrency.lockutils [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.410s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.809824] env[62692]: INFO nova.compute.manager [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Migrating [ 1133.816800] env[62692]: DEBUG oslo_concurrency.lockutils [None req-60d6f9f1-5668-4939-9776-13acb47a0e72 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "1c7ce408-6e9c-43bd-8d81-6acb12353ed8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.219s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.829302] env[62692]: DEBUG oslo_vmware.api [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141904, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.831245] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1133.831467] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1133.831609] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Deleting the datastore file [datastore1] 7183ec4a-992d-48f1-8dda-7f499c2f4e1d {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1133.832380] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bee33315-8c98-472e-816a-db906ce33384 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.840113] env[62692]: DEBUG oslo_vmware.api [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1133.840113] env[62692]: value = "task-1141905" [ 1133.840113] env[62692]: _type = "Task" [ 1133.840113] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.850369] env[62692]: DEBUG oslo_vmware.api [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141905, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.025446] env[62692]: DEBUG oslo_vmware.api [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141896, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.032182] env[62692]: DEBUG oslo_vmware.api [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141901, 'name': PowerOnVM_Task, 'duration_secs': 0.575563} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.032182] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1134.035048] env[62692]: DEBUG nova.compute.manager [None req-a4362d16-2c1b-4563-8ecc-03172af71c5b tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1134.035695] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d3bb84f-e41a-407d-8ed5-8de39ae60405 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.141783] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Releasing lock "refresh_cache-c20139ae-5076-4f6b-b4f0-047332559acf" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1134.142147] env[62692]: DEBUG nova.compute.manager [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Instance network_info: |[{"id": "e93a1d6d-aa66-47ac-8d7c-db90c1d090f8", "address": "fa:16:3e:e7:6a:de", "network": {"id": "647be031-b44d-4439-a059-a40feedbbccf", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-477265227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a331bc93fe1140b68242feba6b02655a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f246b87-f105-4b33-a71d-5caf8e99e074", "external-id": "nsx-vlan-transportzone-583", "segmentation_id": 583, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape93a1d6d-aa", "ovs_interfaceid": "e93a1d6d-aa66-47ac-8d7c-db90c1d090f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1134.142496] env[62692]: DEBUG oslo_concurrency.lockutils [req-37ff6ccd-9e01-411f-b9d8-2db1b92184d8 req-c87f53cf-2cf1-4785-a021-d7664ba82f20 service nova] Acquired lock "refresh_cache-c20139ae-5076-4f6b-b4f0-047332559acf" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.142789] env[62692]: DEBUG nova.network.neutron [req-37ff6ccd-9e01-411f-b9d8-2db1b92184d8 req-c87f53cf-2cf1-4785-a021-d7664ba82f20 service nova] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Refreshing network info cache for port e93a1d6d-aa66-47ac-8d7c-db90c1d090f8 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1134.144457] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e7:6a:de', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0f246b87-f105-4b33-a71d-5caf8e99e074', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e93a1d6d-aa66-47ac-8d7c-db90c1d090f8', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1134.152243] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Creating folder: Project (a331bc93fe1140b68242feba6b02655a). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1134.153550] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0e17cb1c-5370-4fda-9529-c1f76dc7e6c9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.167567] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Created folder: Project (a331bc93fe1140b68242feba6b02655a) in parent group-v248868. [ 1134.167781] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Creating folder: Instances. Parent ref: group-v249064. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1134.168045] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5b806d74-3c0e-4cf9-a338-cc188d82aa92 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.179578] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Created folder: Instances in parent group-v249064. [ 1134.179958] env[62692]: DEBUG oslo.service.loopingcall [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1134.180267] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1134.180811] env[62692]: INFO nova.compute.manager [-] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Took 1.77 seconds to deallocate network for instance. [ 1134.181178] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9e0d1767-a27a-4b6b-a2bc-9a86195aa8ad {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.210594] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1134.210594] env[62692]: value = "task-1141908" [ 1134.210594] env[62692]: _type = "Task" [ 1134.210594] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.219730] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141908, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.318749] env[62692]: DEBUG oslo_vmware.api [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141904, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.325444] env[62692]: DEBUG oslo_concurrency.lockutils [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "refresh_cache-e55a1a5b-3656-48d7-a610-04610b43999a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1134.325653] env[62692]: DEBUG oslo_concurrency.lockutils [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired lock "refresh_cache-e55a1a5b-3656-48d7-a610-04610b43999a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.325835] env[62692]: DEBUG nova.network.neutron [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1134.351070] env[62692]: DEBUG oslo_vmware.api [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141905, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.281569} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.351357] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1134.351569] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1134.351774] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1134.351965] env[62692]: INFO nova.compute.manager [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1134.352248] env[62692]: DEBUG oslo.service.loopingcall [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1134.352471] env[62692]: DEBUG nova.compute.manager [-] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1134.352571] env[62692]: DEBUG nova.network.neutron [-] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1134.524506] env[62692]: DEBUG oslo_vmware.api [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141896, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.705665] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.705962] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.706208] env[62692]: DEBUG nova.objects.instance [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lazy-loading 'resources' on Instance uuid a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1134.725081] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141908, 'name': CreateVM_Task, 'duration_secs': 0.433965} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.725081] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1134.725515] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1134.725689] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.726176] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1134.726444] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7cb50889-03e0-4685-b74f-cdc552019697 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.732746] env[62692]: DEBUG oslo_vmware.api [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Waiting for the task: (returnval){ [ 1134.732746] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52419985-e9ae-c4a7-3e78-8e19f8d55705" [ 1134.732746] env[62692]: _type = "Task" [ 1134.732746] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.743046] env[62692]: DEBUG oslo_vmware.api [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52419985-e9ae-c4a7-3e78-8e19f8d55705, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.820716] env[62692]: DEBUG oslo_vmware.api [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141904, 'name': CreateSnapshot_Task, 'duration_secs': 0.856082} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.821196] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Created Snapshot of the VM instance {{(pid=62692) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1134.822319] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f94b048-b491-4dc5-b694-8ee721dd9612 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.983662] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9586b47d-e303-4f81-b29f-8e9593217a1f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.014754] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Updating instance 'cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43' progress to 0 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1135.023549] env[62692]: DEBUG nova.network.neutron [-] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1135.041020] env[62692]: DEBUG oslo_vmware.api [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141896, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.274509] env[62692]: DEBUG oslo_vmware.api [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52419985-e9ae-c4a7-3e78-8e19f8d55705, 'name': SearchDatastore_Task, 'duration_secs': 0.018444} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.274509] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1135.274509] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1135.274509] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1135.274509] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.274509] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1135.274509] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-534f53c6-45fc-42d3-9483-2e82c8dfdbb3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.274509] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1135.274509] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1135.274509] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d36e764-942f-4266-b585-d0e4dc476d20 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.274509] env[62692]: DEBUG oslo_vmware.api [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Waiting for the task: (returnval){ [ 1135.274509] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]522392d3-5d6c-308c-2b26-194f8468185c" [ 1135.274509] env[62692]: _type = "Task" [ 1135.274509] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.274509] env[62692]: DEBUG oslo_vmware.api [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]522392d3-5d6c-308c-2b26-194f8468185c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.350601] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Creating linked-clone VM from snapshot {{(pid=62692) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1135.358018] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-8ed258ec-29b2-45ed-9cfb-83b33dc569aa {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.370648] env[62692]: DEBUG oslo_vmware.api [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1135.370648] env[62692]: value = "task-1141909" [ 1135.370648] env[62692]: _type = "Task" [ 1135.370648] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.382671] env[62692]: DEBUG oslo_vmware.api [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141909, 'name': CloneVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.431091] env[62692]: DEBUG nova.network.neutron [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Updating instance_info_cache with network_info: [{"id": "5e50f271-31a9-4755-a612-6b971179167a", "address": "fa:16:3e:75:3c:39", "network": {"id": "dd260020-2d74-46ba-aa78-af29bdefa176", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-617179655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "589895ae20394e2fa525707bf7d09f2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e50f271-31", "ovs_interfaceid": "5e50f271-31a9-4755-a612-6b971179167a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1135.532777] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1135.532777] env[62692]: INFO nova.compute.manager [-] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Took 1.18 seconds to deallocate network for instance. [ 1135.532777] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-660f5478-6cb5-4ed5-b14f-4c8eb1dc6c4a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.547421] env[62692]: DEBUG oslo_vmware.api [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141896, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.549868] env[62692]: DEBUG oslo_vmware.api [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1135.549868] env[62692]: value = "task-1141911" [ 1135.549868] env[62692]: _type = "Task" [ 1135.549868] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.564895] env[62692]: DEBUG oslo_vmware.api [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141911, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.566728] env[62692]: DEBUG nova.network.neutron [req-37ff6ccd-9e01-411f-b9d8-2db1b92184d8 req-c87f53cf-2cf1-4785-a021-d7664ba82f20 service nova] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Updated VIF entry in instance network info cache for port e93a1d6d-aa66-47ac-8d7c-db90c1d090f8. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1135.566728] env[62692]: DEBUG nova.network.neutron [req-37ff6ccd-9e01-411f-b9d8-2db1b92184d8 req-c87f53cf-2cf1-4785-a021-d7664ba82f20 service nova] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Updating instance_info_cache with network_info: [{"id": "e93a1d6d-aa66-47ac-8d7c-db90c1d090f8", "address": "fa:16:3e:e7:6a:de", "network": {"id": "647be031-b44d-4439-a059-a40feedbbccf", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-477265227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a331bc93fe1140b68242feba6b02655a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f246b87-f105-4b33-a71d-5caf8e99e074", "external-id": "nsx-vlan-transportzone-583", "segmentation_id": 583, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape93a1d6d-aa", "ovs_interfaceid": "e93a1d6d-aa66-47ac-8d7c-db90c1d090f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1135.571126] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8f440fa-a0e1-4d5c-926d-f8759a83c211 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.580385] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85f7cd85-3ceb-4c28-a74d-1827732d6803 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.616245] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-900584f4-8619-402d-bbf3-e955acebfb2d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.626633] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2415f3d3-1c64-408d-bbb5-66a92db98f9e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.648065] env[62692]: DEBUG nova.compute.provider_tree [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1135.682708] env[62692]: DEBUG nova.compute.manager [req-2d261672-9628-4043-ac4c-ed030469db58 req-60826468-b9e3-43d4-9a09-f71196b83bde service nova] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Received event network-changed-5c943de3-e4d0-479b-9d1d-b38816b891b2 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1135.682786] env[62692]: DEBUG nova.compute.manager [req-2d261672-9628-4043-ac4c-ed030469db58 req-60826468-b9e3-43d4-9a09-f71196b83bde service nova] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Refreshing instance network info cache due to event network-changed-5c943de3-e4d0-479b-9d1d-b38816b891b2. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1135.683021] env[62692]: DEBUG oslo_concurrency.lockutils [req-2d261672-9628-4043-ac4c-ed030469db58 req-60826468-b9e3-43d4-9a09-f71196b83bde service nova] Acquiring lock "refresh_cache-1c7ce408-6e9c-43bd-8d81-6acb12353ed8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1135.683179] env[62692]: DEBUG oslo_concurrency.lockutils [req-2d261672-9628-4043-ac4c-ed030469db58 req-60826468-b9e3-43d4-9a09-f71196b83bde service nova] Acquired lock "refresh_cache-1c7ce408-6e9c-43bd-8d81-6acb12353ed8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.683408] env[62692]: DEBUG nova.network.neutron [req-2d261672-9628-4043-ac4c-ed030469db58 req-60826468-b9e3-43d4-9a09-f71196b83bde service nova] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Refreshing network info cache for port 5c943de3-e4d0-479b-9d1d-b38816b891b2 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1135.775052] env[62692]: DEBUG oslo_vmware.api [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]522392d3-5d6c-308c-2b26-194f8468185c, 'name': SearchDatastore_Task, 'duration_secs': 0.011035} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.775905] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dbd1edcd-d616-47a2-9655-e89609a22ffb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.783669] env[62692]: DEBUG oslo_vmware.api [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Waiting for the task: (returnval){ [ 1135.783669] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5243817a-b4c1-b4e3-25b1-c131f5ed7c65" [ 1135.783669] env[62692]: _type = "Task" [ 1135.783669] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.806322] env[62692]: DEBUG oslo_vmware.api [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5243817a-b4c1-b4e3-25b1-c131f5ed7c65, 'name': SearchDatastore_Task, 'duration_secs': 0.018825} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.806691] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1135.806989] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] c20139ae-5076-4f6b-b4f0-047332559acf/c20139ae-5076-4f6b-b4f0-047332559acf.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1135.807310] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-346b07dd-a29d-42b3-9299-6af331807d12 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.816895] env[62692]: DEBUG oslo_vmware.api [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Waiting for the task: (returnval){ [ 1135.816895] env[62692]: value = "task-1141912" [ 1135.816895] env[62692]: _type = "Task" [ 1135.816895] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.827611] env[62692]: DEBUG oslo_vmware.api [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Task: {'id': task-1141912, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.882396] env[62692]: DEBUG oslo_vmware.api [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141909, 'name': CloneVM_Task} progress is 94%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.934500] env[62692]: DEBUG oslo_concurrency.lockutils [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Releasing lock "refresh_cache-e55a1a5b-3656-48d7-a610-04610b43999a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1136.036458] env[62692]: DEBUG oslo_vmware.api [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141896, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.050087] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.068569] env[62692]: DEBUG oslo_vmware.api [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141911, 'name': PowerOffVM_Task, 'duration_secs': 0.277446} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.069191] env[62692]: DEBUG oslo_concurrency.lockutils [req-37ff6ccd-9e01-411f-b9d8-2db1b92184d8 req-c87f53cf-2cf1-4785-a021-d7664ba82f20 service nova] Releasing lock "refresh_cache-c20139ae-5076-4f6b-b4f0-047332559acf" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1136.069497] env[62692]: DEBUG nova.compute.manager [req-37ff6ccd-9e01-411f-b9d8-2db1b92184d8 req-c87f53cf-2cf1-4785-a021-d7664ba82f20 service nova] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Received event network-vif-deleted-c3aed689-318a-45c3-a236-6ca6b1bace9b {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1136.069756] env[62692]: INFO nova.compute.manager [req-37ff6ccd-9e01-411f-b9d8-2db1b92184d8 req-c87f53cf-2cf1-4785-a021-d7664ba82f20 service nova] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Neutron deleted interface c3aed689-318a-45c3-a236-6ca6b1bace9b; detaching it from the instance and deleting it from the info cache [ 1136.069987] env[62692]: DEBUG nova.network.neutron [req-37ff6ccd-9e01-411f-b9d8-2db1b92184d8 req-c87f53cf-2cf1-4785-a021-d7664ba82f20 service nova] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1136.071554] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1136.071767] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Updating instance 'cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43' progress to 17 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1136.148705] env[62692]: DEBUG nova.scheduler.client.report [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1136.328973] env[62692]: DEBUG oslo_vmware.api [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Task: {'id': task-1141912, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.383664] env[62692]: DEBUG oslo_vmware.api [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141909, 'name': CloneVM_Task} progress is 94%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.412603] env[62692]: DEBUG nova.network.neutron [req-2d261672-9628-4043-ac4c-ed030469db58 req-60826468-b9e3-43d4-9a09-f71196b83bde service nova] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Updated VIF entry in instance network info cache for port 5c943de3-e4d0-479b-9d1d-b38816b891b2. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1136.413055] env[62692]: DEBUG nova.network.neutron [req-2d261672-9628-4043-ac4c-ed030469db58 req-60826468-b9e3-43d4-9a09-f71196b83bde service nova] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Updating instance_info_cache with network_info: [{"id": "5c943de3-e4d0-479b-9d1d-b38816b891b2", "address": "fa:16:3e:54:d8:95", "network": {"id": "2febb660-329d-403c-8c2c-f11c0f5f547c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1806650357-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37921f67664544478cb2568def068b11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c943de3-e4", "ovs_interfaceid": "5c943de3-e4d0-479b-9d1d-b38816b891b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1136.536326] env[62692]: DEBUG oslo_vmware.api [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141896, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.548658] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquiring lock "2b21e8a1-5f5e-4f1f-bae6-512f1213eb48" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.548658] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Lock "2b21e8a1-5f5e-4f1f-bae6-512f1213eb48" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.578908] env[62692]: DEBUG nova.virt.hardware [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1136.579121] env[62692]: DEBUG nova.virt.hardware [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1136.579335] env[62692]: DEBUG nova.virt.hardware [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1136.579586] env[62692]: DEBUG nova.virt.hardware [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1136.579758] env[62692]: DEBUG nova.virt.hardware [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1136.579923] env[62692]: DEBUG nova.virt.hardware [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1136.580170] env[62692]: DEBUG nova.virt.hardware [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1136.580357] env[62692]: DEBUG nova.virt.hardware [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1136.580544] env[62692]: DEBUG nova.virt.hardware [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1136.580716] env[62692]: DEBUG nova.virt.hardware [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1136.580899] env[62692]: DEBUG nova.virt.hardware [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1136.586066] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cf84ecd8-5d44-4f47-9d40-b185b9a762a7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.588714] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0e20122e-5720-4ae2-bffb-bfb2a695d329 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.607236] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa3548b1-0d23-4ab3-9a26-509fc8de47aa {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.619180] env[62692]: DEBUG oslo_vmware.api [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1136.619180] env[62692]: value = "task-1141913" [ 1136.619180] env[62692]: _type = "Task" [ 1136.619180] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.630330] env[62692]: DEBUG oslo_vmware.api [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141913, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.642130] env[62692]: DEBUG nova.compute.manager [req-37ff6ccd-9e01-411f-b9d8-2db1b92184d8 req-c87f53cf-2cf1-4785-a021-d7664ba82f20 service nova] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Detach interface failed, port_id=c3aed689-318a-45c3-a236-6ca6b1bace9b, reason: Instance a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1136.653481] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.947s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.655900] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.606s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.656410] env[62692]: DEBUG nova.objects.instance [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lazy-loading 'resources' on Instance uuid 7183ec4a-992d-48f1-8dda-7f499c2f4e1d {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1136.673467] env[62692]: INFO nova.scheduler.client.report [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Deleted allocations for instance a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970 [ 1136.831399] env[62692]: DEBUG oslo_vmware.api [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Task: {'id': task-1141912, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.550459} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.831835] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] c20139ae-5076-4f6b-b4f0-047332559acf/c20139ae-5076-4f6b-b4f0-047332559acf.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1136.832228] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1136.832613] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c60b6832-7adf-45d1-b68e-81c70672a171 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.842254] env[62692]: DEBUG oslo_vmware.api [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Waiting for the task: (returnval){ [ 1136.842254] env[62692]: value = "task-1141914" [ 1136.842254] env[62692]: _type = "Task" [ 1136.842254] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.855554] env[62692]: DEBUG oslo_vmware.api [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Task: {'id': task-1141914, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.883978] env[62692]: DEBUG oslo_vmware.api [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141909, 'name': CloneVM_Task} progress is 95%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.916194] env[62692]: DEBUG oslo_concurrency.lockutils [req-2d261672-9628-4043-ac4c-ed030469db58 req-60826468-b9e3-43d4-9a09-f71196b83bde service nova] Releasing lock "refresh_cache-1c7ce408-6e9c-43bd-8d81-6acb12353ed8" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1136.916520] env[62692]: DEBUG nova.compute.manager [req-2d261672-9628-4043-ac4c-ed030469db58 req-60826468-b9e3-43d4-9a09-f71196b83bde service nova] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Received event network-vif-deleted-ab9d5209-bc8f-437d-a219-7f521cae10e6 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1137.037191] env[62692]: DEBUG oslo_vmware.api [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141896, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.050843] env[62692]: DEBUG nova.compute.manager [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1137.131888] env[62692]: DEBUG oslo_vmware.api [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141913, 'name': ReconfigVM_Task, 'duration_secs': 0.393452} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.132165] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Updating instance 'cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43' progress to 33 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1137.182300] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e0601bac-336b-4865-9b62-5b213ad0a88e tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.932s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.350426] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acca6106-2c68-47b7-8e2e-6c59a2c1c699 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.365689] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e2271b4-5399-42a4-86b0-8bd86ee62054 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.370775] env[62692]: DEBUG oslo_vmware.api [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Task: {'id': task-1141914, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.18735} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.371183] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1137.372779] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90791709-c0de-433e-a136-d1a5506b846a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.424095] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce6d813e-c90f-4065-85af-56afcd30c688 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.449921] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] c20139ae-5076-4f6b-b4f0-047332559acf/c20139ae-5076-4f6b-b4f0-047332559acf.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1137.453362] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f39555c4-aad3-43fc-91d3-7df0194dfe79 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.477063] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b63402c3-090a-40e5-9d81-353bb14dbaa9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.479833] env[62692]: DEBUG oslo_vmware.api [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141909, 'name': CloneVM_Task, 'duration_secs': 1.77372} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.481519] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bf36537-0f91-4623-b745-948a17cbcecf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.485938] env[62692]: INFO nova.virt.vmwareapi.vmops [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Created linked-clone VM from snapshot [ 1137.488509] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ce30c3a-b907-4350-aa60-657f6ec9f0a7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.491365] env[62692]: DEBUG oslo_vmware.api [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Waiting for the task: (returnval){ [ 1137.491365] env[62692]: value = "task-1141915" [ 1137.491365] env[62692]: _type = "Task" [ 1137.491365] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.506063] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Updating instance 'e55a1a5b-3656-48d7-a610-04610b43999a' progress to 0 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1137.522181] env[62692]: DEBUG nova.compute.provider_tree [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1137.527229] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Uploading image 59f3345e-2371-4b7f-9944-01f2adf5a997 {{(pid=62692) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1137.536568] env[62692]: DEBUG oslo_vmware.api [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Task: {'id': task-1141915, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.544788] env[62692]: DEBUG oslo_vmware.api [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141896, 'name': ReconfigVM_Task, 'duration_secs': 5.936479} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.544956] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1137.545160] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Reconfigured VM to detach interface {{(pid=62692) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1137.561440] env[62692]: DEBUG oslo_vmware.rw_handles [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1137.561440] env[62692]: value = "vm-249068" [ 1137.561440] env[62692]: _type = "VirtualMachine" [ 1137.561440] env[62692]: }. {{(pid=62692) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1137.561785] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-9717d2b2-48d0-4965-b72d-0c5bbcfab68f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.570811] env[62692]: DEBUG oslo_vmware.rw_handles [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lease: (returnval){ [ 1137.570811] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52712929-3533-af44-5503-f3de1ca553e5" [ 1137.570811] env[62692]: _type = "HttpNfcLease" [ 1137.570811] env[62692]: } obtained for exporting VM: (result){ [ 1137.570811] env[62692]: value = "vm-249068" [ 1137.570811] env[62692]: _type = "VirtualMachine" [ 1137.570811] env[62692]: }. {{(pid=62692) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1137.571122] env[62692]: DEBUG oslo_vmware.api [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the lease: (returnval){ [ 1137.571122] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52712929-3533-af44-5503-f3de1ca553e5" [ 1137.571122] env[62692]: _type = "HttpNfcLease" [ 1137.571122] env[62692]: } to be ready. {{(pid=62692) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1137.575260] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.582083] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1137.582083] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52712929-3533-af44-5503-f3de1ca553e5" [ 1137.582083] env[62692]: _type = "HttpNfcLease" [ 1137.582083] env[62692]: } is initializing. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1137.639021] env[62692]: DEBUG nova.virt.hardware [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1137.639310] env[62692]: DEBUG nova.virt.hardware [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1137.639648] env[62692]: DEBUG nova.virt.hardware [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1137.639648] env[62692]: DEBUG nova.virt.hardware [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1137.639903] env[62692]: DEBUG nova.virt.hardware [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1137.640048] env[62692]: DEBUG nova.virt.hardware [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1137.640223] env[62692]: DEBUG nova.virt.hardware [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1137.640393] env[62692]: DEBUG nova.virt.hardware [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1137.640889] env[62692]: DEBUG nova.virt.hardware [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1137.640889] env[62692]: DEBUG nova.virt.hardware [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1137.640889] env[62692]: DEBUG nova.virt.hardware [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1137.647023] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Reconfiguring VM instance instance-00000062 to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1137.647432] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9350ed20-cff6-4aa7-9c2b-b4d60f616e01 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.671068] env[62692]: DEBUG oslo_vmware.api [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1137.671068] env[62692]: value = "task-1141917" [ 1137.671068] env[62692]: _type = "Task" [ 1137.671068] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.681451] env[62692]: DEBUG oslo_vmware.api [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141917, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.017227] env[62692]: DEBUG oslo_vmware.api [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Task: {'id': task-1141915, 'name': ReconfigVM_Task, 'duration_secs': 0.446873} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.017585] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Reconfigured VM instance instance-0000006c to attach disk [datastore2] c20139ae-5076-4f6b-b4f0-047332559acf/c20139ae-5076-4f6b-b4f0-047332559acf.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1138.018307] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-faca583e-be46-4f97-ba7e-8bdd999e251a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.029384] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1138.031262] env[62692]: DEBUG nova.scheduler.client.report [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1138.035815] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-73c7c92e-abf7-47f3-9d05-e30234c2d92f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.041023] env[62692]: DEBUG oslo_vmware.api [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Waiting for the task: (returnval){ [ 1138.041023] env[62692]: value = "task-1141918" [ 1138.041023] env[62692]: _type = "Task" [ 1138.041023] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.052075] env[62692]: DEBUG oslo_vmware.api [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Task: {'id': task-1141918, 'name': Rename_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.053099] env[62692]: DEBUG oslo_vmware.api [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1138.053099] env[62692]: value = "task-1141919" [ 1138.053099] env[62692]: _type = "Task" [ 1138.053099] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.068676] env[62692]: DEBUG oslo_vmware.api [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141919, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.085238] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1138.085238] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52712929-3533-af44-5503-f3de1ca553e5" [ 1138.085238] env[62692]: _type = "HttpNfcLease" [ 1138.085238] env[62692]: } is ready. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1138.085635] env[62692]: DEBUG oslo_vmware.rw_handles [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1138.085635] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52712929-3533-af44-5503-f3de1ca553e5" [ 1138.085635] env[62692]: _type = "HttpNfcLease" [ 1138.085635] env[62692]: }. {{(pid=62692) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1138.086706] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95ab0f98-a236-4ff8-ad7a-8eff3a55eed0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.100696] env[62692]: DEBUG oslo_vmware.rw_handles [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526094c2-bec6-23ad-7aed-b16318e5bb13/disk-0.vmdk from lease info. {{(pid=62692) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1138.101787] env[62692]: DEBUG oslo_vmware.rw_handles [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526094c2-bec6-23ad-7aed-b16318e5bb13/disk-0.vmdk for reading. {{(pid=62692) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1138.183678] env[62692]: DEBUG oslo_vmware.api [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141917, 'name': ReconfigVM_Task, 'duration_secs': 0.202773} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.184202] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Reconfigured VM instance instance-00000062 to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1138.185441] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0e76969-aa47-433d-83e0-7ea08ed71bdb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.210299] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43/cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1138.213064] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-71911150-e6a3-4cf0-bfea-fc58a05d04e3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.227167] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b6aa0e26-b4f8-46c7-85e4-bd4d3a1a47db {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.238453] env[62692]: DEBUG oslo_vmware.api [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1138.238453] env[62692]: value = "task-1141920" [ 1138.238453] env[62692]: _type = "Task" [ 1138.238453] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.251094] env[62692]: DEBUG oslo_vmware.api [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141920, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.539881] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.884s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.542524] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.967s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.544170] env[62692]: INFO nova.compute.claims [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1138.564268] env[62692]: DEBUG oslo_vmware.api [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Task: {'id': task-1141918, 'name': Rename_Task, 'duration_secs': 0.270928} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.565271] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1138.565563] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4aedf30a-cdbe-4305-8950-e120534f2638 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.573020] env[62692]: DEBUG oslo_vmware.api [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141919, 'name': PowerOffVM_Task, 'duration_secs': 0.310902} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.574093] env[62692]: INFO nova.scheduler.client.report [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Deleted allocations for instance 7183ec4a-992d-48f1-8dda-7f499c2f4e1d [ 1138.575503] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1138.575694] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Updating instance 'e55a1a5b-3656-48d7-a610-04610b43999a' progress to 17 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1138.583215] env[62692]: DEBUG oslo_vmware.api [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Waiting for the task: (returnval){ [ 1138.583215] env[62692]: value = "task-1141922" [ 1138.583215] env[62692]: _type = "Task" [ 1138.583215] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.595834] env[62692]: DEBUG oslo_vmware.api [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Task: {'id': task-1141922, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.749897] env[62692]: DEBUG oslo_vmware.api [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141920, 'name': ReconfigVM_Task, 'duration_secs': 0.446983} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.750499] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Reconfigured VM instance instance-00000062 to attach disk [datastore2] cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43/cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1138.751104] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Updating instance 'cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43' progress to 50 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1139.004338] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "refresh_cache-94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1139.004606] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "refresh_cache-94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1139.004875] env[62692]: DEBUG nova.network.neutron [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1139.084759] env[62692]: DEBUG nova.virt.hardware [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1139.085414] env[62692]: DEBUG nova.virt.hardware [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1139.085641] env[62692]: DEBUG nova.virt.hardware [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1139.085858] env[62692]: DEBUG nova.virt.hardware [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1139.086048] env[62692]: DEBUG nova.virt.hardware [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1139.086303] env[62692]: DEBUG nova.virt.hardware [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1139.086605] env[62692]: DEBUG nova.virt.hardware [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1139.086817] env[62692]: DEBUG nova.virt.hardware [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1139.087031] env[62692]: DEBUG nova.virt.hardware [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1139.087497] env[62692]: DEBUG nova.virt.hardware [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1139.087830] env[62692]: DEBUG nova.virt.hardware [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1139.098181] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-583cbe65-91ac-4dcf-bb65-e01e9312d1ba {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.114087] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8f35345b-88ed-4f36-92ed-df34cf9a4ff6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "7183ec4a-992d-48f1-8dda-7f499c2f4e1d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.902s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.135275] env[62692]: DEBUG oslo_vmware.api [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Task: {'id': task-1141922, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.137088] env[62692]: DEBUG oslo_vmware.api [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1139.137088] env[62692]: value = "task-1141923" [ 1139.137088] env[62692]: _type = "Task" [ 1139.137088] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.150524] env[62692]: DEBUG oslo_vmware.api [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141923, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.262596] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54e839b1-07ef-4e34-ad12-8828468c9545 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.286589] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d99d391e-42e0-4935-af93-23f3acc45b85 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.308500] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Updating instance 'cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43' progress to 67 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1139.612707] env[62692]: DEBUG oslo_vmware.api [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Task: {'id': task-1141922, 'name': PowerOnVM_Task, 'duration_secs': 0.92634} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.612930] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1139.613056] env[62692]: INFO nova.compute.manager [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Took 8.56 seconds to spawn the instance on the hypervisor. [ 1139.613381] env[62692]: DEBUG nova.compute.manager [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1139.614484] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91699d4d-fcbb-4105-8dae-6c7d8e415efc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.654878] env[62692]: DEBUG oslo_vmware.api [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141923, 'name': ReconfigVM_Task, 'duration_secs': 0.332626} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.655541] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Updating instance 'e55a1a5b-3656-48d7-a610-04610b43999a' progress to 33 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1139.802440] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3adfe1ab-b323-4a8a-8b1b-9d14008e498f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.811908] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cedb699-28f7-46c4-9621-8041a34d4bdb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.822292] env[62692]: INFO nova.network.neutron [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Port 613360e6-9cd1-4423-9c92-293618e0c0d1 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1139.822782] env[62692]: DEBUG nova.network.neutron [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Updating instance_info_cache with network_info: [{"id": "d51ffd3f-c023-45e2-8bfa-1424b7ccb188", "address": "fa:16:3e:a8:44:f1", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd51ffd3f-c0", "ovs_interfaceid": "d51ffd3f-c023-45e2-8bfa-1424b7ccb188", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.860870] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-170ecfd0-c005-4c05-a64a-f4625bdad71c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.870425] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93ad4ad9-815c-4d53-bdd1-829337300bc5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.887720] env[62692]: DEBUG nova.compute.provider_tree [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1139.893021] env[62692]: DEBUG nova.network.neutron [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Port a3885c51-6f7b-41e8-8948-41fac442a288 binding to destination host cpu-1 is already ACTIVE {{(pid=62692) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1140.145078] env[62692]: INFO nova.compute.manager [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Took 13.49 seconds to build instance. [ 1140.156305] env[62692]: DEBUG nova.compute.manager [req-977d33c9-c612-4296-8a25-fce3fb47d165 req-4be24368-95c9-4b7c-b157-e8dbdc9212ef service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Received event network-changed-d51ffd3f-c023-45e2-8bfa-1424b7ccb188 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1140.156631] env[62692]: DEBUG nova.compute.manager [req-977d33c9-c612-4296-8a25-fce3fb47d165 req-4be24368-95c9-4b7c-b157-e8dbdc9212ef service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Refreshing instance network info cache due to event network-changed-d51ffd3f-c023-45e2-8bfa-1424b7ccb188. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1140.158958] env[62692]: DEBUG oslo_concurrency.lockutils [req-977d33c9-c612-4296-8a25-fce3fb47d165 req-4be24368-95c9-4b7c-b157-e8dbdc9212ef service nova] Acquiring lock "refresh_cache-94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1140.167094] env[62692]: DEBUG nova.virt.hardware [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1140.167815] env[62692]: DEBUG nova.virt.hardware [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1140.167815] env[62692]: DEBUG nova.virt.hardware [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1140.168164] env[62692]: DEBUG nova.virt.hardware [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1140.168164] env[62692]: DEBUG nova.virt.hardware [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1140.168259] env[62692]: DEBUG nova.virt.hardware [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1140.168418] env[62692]: DEBUG nova.virt.hardware [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1140.168636] env[62692]: DEBUG nova.virt.hardware [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1140.168868] env[62692]: DEBUG nova.virt.hardware [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1140.169186] env[62692]: DEBUG nova.virt.hardware [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1140.169427] env[62692]: DEBUG nova.virt.hardware [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1140.174979] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Reconfiguring VM instance instance-00000069 to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1140.176534] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-76072a07-8f28-49da-85be-ea25422e11b1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.200013] env[62692]: DEBUG oslo_vmware.api [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1140.200013] env[62692]: value = "task-1141924" [ 1140.200013] env[62692]: _type = "Task" [ 1140.200013] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.209909] env[62692]: DEBUG oslo_vmware.api [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141924, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.258699] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "0b7ea87c-252a-48b3-a1cd-c81e91362812" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.258967] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "0b7ea87c-252a-48b3-a1cd-c81e91362812" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.326233] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "refresh_cache-94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1140.332017] env[62692]: DEBUG oslo_concurrency.lockutils [req-977d33c9-c612-4296-8a25-fce3fb47d165 req-4be24368-95c9-4b7c-b157-e8dbdc9212ef service nova] Acquired lock "refresh_cache-94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.332017] env[62692]: DEBUG nova.network.neutron [req-977d33c9-c612-4296-8a25-fce3fb47d165 req-4be24368-95c9-4b7c-b157-e8dbdc9212ef service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Refreshing network info cache for port d51ffd3f-c023-45e2-8bfa-1424b7ccb188 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1140.395826] env[62692]: DEBUG nova.scheduler.client.report [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1140.647497] env[62692]: DEBUG oslo_concurrency.lockutils [None req-51c3d2db-6420-47ed-9402-211029b4cb92 tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Lock "c20139ae-5076-4f6b-b4f0-047332559acf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.997s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.710772] env[62692]: DEBUG oslo_vmware.api [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141924, 'name': ReconfigVM_Task, 'duration_secs': 0.290795} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.711120] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Reconfigured VM instance instance-00000069 to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1140.711979] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c8a9065-caa5-445f-931c-ff340c4f1037 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.735973] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] e55a1a5b-3656-48d7-a610-04610b43999a/e55a1a5b-3656-48d7-a610-04610b43999a.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1140.736391] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9d3372be-7d21-4d08-81d1-d87320fd5634 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.756577] env[62692]: DEBUG oslo_vmware.api [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1140.756577] env[62692]: value = "task-1141926" [ 1140.756577] env[62692]: _type = "Task" [ 1140.756577] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.764602] env[62692]: DEBUG nova.compute.manager [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1140.768647] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Acquiring lock "c20139ae-5076-4f6b-b4f0-047332559acf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.769551] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Lock "c20139ae-5076-4f6b-b4f0-047332559acf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.769927] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Acquiring lock "c20139ae-5076-4f6b-b4f0-047332559acf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.770164] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Lock "c20139ae-5076-4f6b-b4f0-047332559acf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.770372] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Lock "c20139ae-5076-4f6b-b4f0-047332559acf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.772667] env[62692]: INFO nova.compute.manager [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Terminating instance [ 1140.774762] env[62692]: DEBUG nova.compute.manager [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1140.774985] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1140.779203] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b3d8a41-c996-4b42-be50-b54337499d32 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.782408] env[62692]: DEBUG oslo_vmware.api [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141926, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.789362] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1140.789925] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-36f8e558-de5f-4b8a-9717-f27255e00c27 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.801899] env[62692]: DEBUG oslo_vmware.api [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Waiting for the task: (returnval){ [ 1140.801899] env[62692]: value = "task-1141927" [ 1140.801899] env[62692]: _type = "Task" [ 1140.801899] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.814866] env[62692]: DEBUG oslo_vmware.api [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Task: {'id': task-1141927, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.832808] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f11d4f5a-11a7-47fb-a839-a34a8913052c tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "interface-94a07505-661f-4595-9024-8a9da9f2c2fc-613360e6-9cd1-4423-9c92-293618e0c0d1" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.912s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.900094] env[62692]: DEBUG oslo_concurrency.lockutils [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "interface-1904edb2-fd72-43b6-9219-8043c5abff0c-613360e6-9cd1-4423-9c92-293618e0c0d1" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.900610] env[62692]: DEBUG oslo_concurrency.lockutils [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "interface-1904edb2-fd72-43b6-9219-8043c5abff0c-613360e6-9cd1-4423-9c92-293618e0c0d1" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.901010] env[62692]: DEBUG nova.objects.instance [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lazy-loading 'flavor' on Instance uuid 1904edb2-fd72-43b6-9219-8043c5abff0c {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1140.904570] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.362s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.905104] env[62692]: DEBUG nova.compute.manager [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1140.928764] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.929130] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.929333] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.105377] env[62692]: DEBUG nova.network.neutron [req-977d33c9-c612-4296-8a25-fce3fb47d165 req-4be24368-95c9-4b7c-b157-e8dbdc9212ef service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Updated VIF entry in instance network info cache for port d51ffd3f-c023-45e2-8bfa-1424b7ccb188. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1141.105833] env[62692]: DEBUG nova.network.neutron [req-977d33c9-c612-4296-8a25-fce3fb47d165 req-4be24368-95c9-4b7c-b157-e8dbdc9212ef service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Updating instance_info_cache with network_info: [{"id": "d51ffd3f-c023-45e2-8bfa-1424b7ccb188", "address": "fa:16:3e:a8:44:f1", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd51ffd3f-c0", "ovs_interfaceid": "d51ffd3f-c023-45e2-8bfa-1424b7ccb188", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1141.273233] env[62692]: DEBUG oslo_vmware.api [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141926, 'name': ReconfigVM_Task, 'duration_secs': 0.402652} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.276118] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Reconfigured VM instance instance-00000069 to attach disk [datastore1] e55a1a5b-3656-48d7-a610-04610b43999a/e55a1a5b-3656-48d7-a610-04610b43999a.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1141.276118] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Updating instance 'e55a1a5b-3656-48d7-a610-04610b43999a' progress to 50 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1141.291850] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.292151] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.293948] env[62692]: INFO nova.compute.claims [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1141.316029] env[62692]: DEBUG oslo_vmware.api [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Task: {'id': task-1141927, 'name': PowerOffVM_Task, 'duration_secs': 0.194191} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.316029] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1141.316029] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1141.316029] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9bb1f36c-0919-41c2-94db-05bf21f74b73 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.389653] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1141.389899] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1141.390115] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Deleting the datastore file [datastore2] c20139ae-5076-4f6b-b4f0-047332559acf {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1141.390392] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fafb5db2-d705-402a-b17e-ba404b040c41 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.399127] env[62692]: DEBUG oslo_vmware.api [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Waiting for the task: (returnval){ [ 1141.399127] env[62692]: value = "task-1141929" [ 1141.399127] env[62692]: _type = "Task" [ 1141.399127] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.409384] env[62692]: DEBUG oslo_vmware.api [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Task: {'id': task-1141929, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.423488] env[62692]: DEBUG nova.compute.utils [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1141.425000] env[62692]: DEBUG nova.compute.manager [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1141.425185] env[62692]: DEBUG nova.network.neutron [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1141.445162] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "45eccffc-f059-4bef-97da-dacb16f20f88" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.445419] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "45eccffc-f059-4bef-97da-dacb16f20f88" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.497702] env[62692]: DEBUG nova.policy [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6993a074d1384d1e8912e056f0a4caef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b1c44eb72b334ee3add29026d312e37b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1141.512018] env[62692]: DEBUG nova.objects.instance [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lazy-loading 'pci_requests' on Instance uuid 1904edb2-fd72-43b6-9219-8043c5abff0c {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1141.609357] env[62692]: DEBUG oslo_concurrency.lockutils [req-977d33c9-c612-4296-8a25-fce3fb47d165 req-4be24368-95c9-4b7c-b157-e8dbdc9212ef service nova] Releasing lock "refresh_cache-94a07505-661f-4595-9024-8a9da9f2c2fc" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1141.783108] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce64054d-cf9e-4d78-8af1-2fb088bb563f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.809119] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84781d16-c916-4380-ad17-f574f91f7405 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.827948] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Updating instance 'e55a1a5b-3656-48d7-a610-04610b43999a' progress to 67 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1141.851732] env[62692]: DEBUG nova.network.neutron [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Successfully created port: 144e96aa-d02a-48e4-a91e-73909db5c202 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1141.913948] env[62692]: DEBUG oslo_vmware.api [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Task: {'id': task-1141929, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.39061} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.914404] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1141.914864] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1141.915031] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1141.915771] env[62692]: INFO nova.compute.manager [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1141.915771] env[62692]: DEBUG oslo.service.loopingcall [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1141.916039] env[62692]: DEBUG nova.compute.manager [-] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1141.916181] env[62692]: DEBUG nova.network.neutron [-] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1141.928287] env[62692]: DEBUG nova.compute.manager [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1141.948935] env[62692]: DEBUG nova.compute.manager [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1141.992169] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "refresh_cache-cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1141.992369] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired lock "refresh_cache-cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.992558] env[62692]: DEBUG nova.network.neutron [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1142.014505] env[62692]: DEBUG nova.objects.base [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Object Instance<1904edb2-fd72-43b6-9219-8043c5abff0c> lazy-loaded attributes: flavor,pci_requests {{(pid=62692) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1142.014692] env[62692]: DEBUG nova.network.neutron [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1142.120867] env[62692]: DEBUG nova.policy [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a251ef5f6cb143178928babc75a0fef0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ba97a775e99a463c845518b0df8cd3e2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1142.187043] env[62692]: DEBUG nova.compute.manager [req-91b412a7-45a4-4ac3-b943-4bbb350a74f2 req-82498ca3-f7e1-40cf-b249-a7f8f5eb1b22 service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Received event network-changed-a668d6b6-4111-4a94-8bc8-09535b29f895 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1142.187240] env[62692]: DEBUG nova.compute.manager [req-91b412a7-45a4-4ac3-b943-4bbb350a74f2 req-82498ca3-f7e1-40cf-b249-a7f8f5eb1b22 service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Refreshing instance network info cache due to event network-changed-a668d6b6-4111-4a94-8bc8-09535b29f895. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1142.187462] env[62692]: DEBUG oslo_concurrency.lockutils [req-91b412a7-45a4-4ac3-b943-4bbb350a74f2 req-82498ca3-f7e1-40cf-b249-a7f8f5eb1b22 service nova] Acquiring lock "refresh_cache-1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1142.187609] env[62692]: DEBUG oslo_concurrency.lockutils [req-91b412a7-45a4-4ac3-b943-4bbb350a74f2 req-82498ca3-f7e1-40cf-b249-a7f8f5eb1b22 service nova] Acquired lock "refresh_cache-1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1142.187777] env[62692]: DEBUG nova.network.neutron [req-91b412a7-45a4-4ac3-b943-4bbb350a74f2 req-82498ca3-f7e1-40cf-b249-a7f8f5eb1b22 service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Refreshing network info cache for port a668d6b6-4111-4a94-8bc8-09535b29f895 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1142.389500] env[62692]: DEBUG nova.network.neutron [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Port 5e50f271-31a9-4755-a612-6b971179167a binding to destination host cpu-1 is already ACTIVE {{(pid=62692) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1142.478532] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.546797] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1f4b0b4-9368-49e0-b6ab-25f7324673ac {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.555845] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7090bb81-ae60-4ef0-baf1-d0757a0c930e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.589077] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3951bc9a-7689-47d7-a9b9-f23256eff360 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.597608] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-472c7390-bbec-4412-876c-f64853239a5a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.613748] env[62692]: DEBUG nova.compute.provider_tree [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1142.678508] env[62692]: DEBUG nova.network.neutron [-] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1142.938680] env[62692]: DEBUG nova.compute.manager [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1142.966677] env[62692]: DEBUG nova.virt.hardware [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1142.966936] env[62692]: DEBUG nova.virt.hardware [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1142.967119] env[62692]: DEBUG nova.virt.hardware [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1142.967311] env[62692]: DEBUG nova.virt.hardware [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1142.967465] env[62692]: DEBUG nova.virt.hardware [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1142.967617] env[62692]: DEBUG nova.virt.hardware [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1142.967842] env[62692]: DEBUG nova.virt.hardware [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1142.968012] env[62692]: DEBUG nova.virt.hardware [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1142.968320] env[62692]: DEBUG nova.virt.hardware [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1142.968529] env[62692]: DEBUG nova.virt.hardware [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1142.968816] env[62692]: DEBUG nova.virt.hardware [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1142.969853] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce77d0aa-5d25-47e3-aeab-99ebf3ba8fbe {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.980940] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f005c919-334e-44f1-b376-d140c9801a17 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.015474] env[62692]: DEBUG nova.network.neutron [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Updating instance_info_cache with network_info: [{"id": "a3885c51-6f7b-41e8-8948-41fac442a288", "address": "fa:16:3e:e4:9b:6f", "network": {"id": "f70727d6-a9aa-44b0-a244-2533b3f96571", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-284105837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1037e5d436bd429391159b87a712f6e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3885c51-6f", "ovs_interfaceid": "a3885c51-6f7b-41e8-8948-41fac442a288", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1143.117353] env[62692]: DEBUG nova.scheduler.client.report [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1143.181520] env[62692]: INFO nova.compute.manager [-] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Took 1.27 seconds to deallocate network for instance. [ 1143.206865] env[62692]: DEBUG nova.network.neutron [req-91b412a7-45a4-4ac3-b943-4bbb350a74f2 req-82498ca3-f7e1-40cf-b249-a7f8f5eb1b22 service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Updated VIF entry in instance network info cache for port a668d6b6-4111-4a94-8bc8-09535b29f895. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1143.207310] env[62692]: DEBUG nova.network.neutron [req-91b412a7-45a4-4ac3-b943-4bbb350a74f2 req-82498ca3-f7e1-40cf-b249-a7f8f5eb1b22 service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Updating instance_info_cache with network_info: [{"id": "a668d6b6-4111-4a94-8bc8-09535b29f895", "address": "fa:16:3e:02:7f:2a", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa668d6b6-41", "ovs_interfaceid": "a668d6b6-4111-4a94-8bc8-09535b29f895", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1143.409129] env[62692]: DEBUG oslo_concurrency.lockutils [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "e55a1a5b-3656-48d7-a610-04610b43999a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.409398] env[62692]: DEBUG oslo_concurrency.lockutils [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "e55a1a5b-3656-48d7-a610-04610b43999a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.409595] env[62692]: DEBUG oslo_concurrency.lockutils [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "e55a1a5b-3656-48d7-a610-04610b43999a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.469358] env[62692]: DEBUG nova.compute.manager [req-6c8d73ac-7d52-453c-b31f-e87b6d49884a req-9265a85b-ae9e-43f9-afd2-9c19f25ea2f5 service nova] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Received event network-vif-plugged-144e96aa-d02a-48e4-a91e-73909db5c202 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1143.470872] env[62692]: DEBUG oslo_concurrency.lockutils [req-6c8d73ac-7d52-453c-b31f-e87b6d49884a req-9265a85b-ae9e-43f9-afd2-9c19f25ea2f5 service nova] Acquiring lock "2b21e8a1-5f5e-4f1f-bae6-512f1213eb48-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.470872] env[62692]: DEBUG oslo_concurrency.lockutils [req-6c8d73ac-7d52-453c-b31f-e87b6d49884a req-9265a85b-ae9e-43f9-afd2-9c19f25ea2f5 service nova] Lock "2b21e8a1-5f5e-4f1f-bae6-512f1213eb48-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.470872] env[62692]: DEBUG oslo_concurrency.lockutils [req-6c8d73ac-7d52-453c-b31f-e87b6d49884a req-9265a85b-ae9e-43f9-afd2-9c19f25ea2f5 service nova] Lock "2b21e8a1-5f5e-4f1f-bae6-512f1213eb48-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.470872] env[62692]: DEBUG nova.compute.manager [req-6c8d73ac-7d52-453c-b31f-e87b6d49884a req-9265a85b-ae9e-43f9-afd2-9c19f25ea2f5 service nova] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] No waiting events found dispatching network-vif-plugged-144e96aa-d02a-48e4-a91e-73909db5c202 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1143.470872] env[62692]: WARNING nova.compute.manager [req-6c8d73ac-7d52-453c-b31f-e87b6d49884a req-9265a85b-ae9e-43f9-afd2-9c19f25ea2f5 service nova] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Received unexpected event network-vif-plugged-144e96aa-d02a-48e4-a91e-73909db5c202 for instance with vm_state building and task_state spawning. [ 1143.518828] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Releasing lock "refresh_cache-cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1143.625056] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.333s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.625627] env[62692]: DEBUG nova.compute.manager [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1143.628442] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.150s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.630121] env[62692]: INFO nova.compute.claims [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1143.670051] env[62692]: DEBUG nova.network.neutron [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Successfully updated port: 144e96aa-d02a-48e4-a91e-73909db5c202 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1143.689806] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.710223] env[62692]: DEBUG oslo_concurrency.lockutils [req-91b412a7-45a4-4ac3-b943-4bbb350a74f2 req-82498ca3-f7e1-40cf-b249-a7f8f5eb1b22 service nova] Releasing lock "refresh_cache-1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1143.940851] env[62692]: DEBUG nova.network.neutron [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Successfully updated port: 613360e6-9cd1-4423-9c92-293618e0c0d1 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1144.043748] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e95dd48-eb85-4c2c-9eec-026df8ba1d20 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.064184] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b8bbb80-e501-4ac0-8f85-493447e33c61 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.072528] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Updating instance 'cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43' progress to 83 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1144.142042] env[62692]: DEBUG nova.compute.utils [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1144.142042] env[62692]: DEBUG nova.compute.manager [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1144.142042] env[62692]: DEBUG nova.network.neutron [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1144.171744] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquiring lock "refresh_cache-2b21e8a1-5f5e-4f1f-bae6-512f1213eb48" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1144.172313] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquired lock "refresh_cache-2b21e8a1-5f5e-4f1f-bae6-512f1213eb48" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.172313] env[62692]: DEBUG nova.network.neutron [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1144.195931] env[62692]: DEBUG nova.policy [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b0c1314a9aed42e5b54a33de96f3db1f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '71b00c566cea4b21811ea1e1ef5ba0f3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1144.214700] env[62692]: DEBUG nova.compute.manager [req-ab4903aa-1d43-4349-b281-0e1412ee39c1 req-672ae230-99ea-4c67-99ce-6e475da26318 service nova] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Received event network-vif-deleted-e93a1d6d-aa66-47ac-8d7c-db90c1d090f8 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1144.444116] env[62692]: DEBUG oslo_concurrency.lockutils [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "refresh_cache-1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1144.444415] env[62692]: DEBUG oslo_concurrency.lockutils [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "refresh_cache-1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.444686] env[62692]: DEBUG nova.network.neutron [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1144.467676] env[62692]: DEBUG oslo_concurrency.lockutils [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "refresh_cache-e55a1a5b-3656-48d7-a610-04610b43999a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1144.467983] env[62692]: DEBUG oslo_concurrency.lockutils [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired lock "refresh_cache-e55a1a5b-3656-48d7-a610-04610b43999a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.468273] env[62692]: DEBUG nova.network.neutron [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1144.513692] env[62692]: DEBUG nova.network.neutron [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Successfully created port: 34de5e92-0920-4a35-bb28-51c7d09bf35d {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1144.579167] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1144.579607] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7d32a5f4-674b-4cc9-a90b-e9b1d8a0ef4c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.588838] env[62692]: DEBUG oslo_vmware.api [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1144.588838] env[62692]: value = "task-1141932" [ 1144.588838] env[62692]: _type = "Task" [ 1144.588838] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.598465] env[62692]: DEBUG oslo_vmware.api [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141932, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.646557] env[62692]: DEBUG nova.compute.manager [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1144.720822] env[62692]: DEBUG nova.network.neutron [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1144.990032] env[62692]: DEBUG nova.network.neutron [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Updating instance_info_cache with network_info: [{"id": "144e96aa-d02a-48e4-a91e-73909db5c202", "address": "fa:16:3e:23:7e:07", "network": {"id": "50c1330e-1a5f-42d1-9987-79ddcc0ad7aa", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-813901268-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b1c44eb72b334ee3add29026d312e37b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f92f0b92-d6fb-4d00-8ad5-6b3809ed5493", "external-id": "nsx-vlan-transportzone-48", "segmentation_id": 48, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap144e96aa-d0", "ovs_interfaceid": "144e96aa-d02a-48e4-a91e-73909db5c202", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1145.001768] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c1a7d03-cafb-4283-b8ba-88b16ec4da3a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.012950] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dacdacb3-d165-4e28-a344-6ce023933a29 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.018518] env[62692]: WARNING nova.network.neutron [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] b34af32c-59b7-4de4-85ef-08bd67c1e2a2 already exists in list: networks containing: ['b34af32c-59b7-4de4-85ef-08bd67c1e2a2']. ignoring it [ 1145.051448] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-108a9100-db11-4c78-ac53-9db698289bbf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.062806] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1056a27b-d4db-4d51-b478-2319f95c22ab {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.079761] env[62692]: DEBUG nova.compute.provider_tree [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1145.100305] env[62692]: DEBUG oslo_vmware.api [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141932, 'name': PowerOnVM_Task, 'duration_secs': 0.436887} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.100599] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1145.100790] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2a640244-940d-4086-9e6c-900d8d12a242 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Updating instance 'cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43' progress to 100 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1145.400975] env[62692]: DEBUG nova.network.neutron [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Updating instance_info_cache with network_info: [{"id": "5e50f271-31a9-4755-a612-6b971179167a", "address": "fa:16:3e:75:3c:39", "network": {"id": "dd260020-2d74-46ba-aa78-af29bdefa176", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-617179655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "589895ae20394e2fa525707bf7d09f2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e50f271-31", "ovs_interfaceid": "5e50f271-31a9-4755-a612-6b971179167a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1145.491432] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Releasing lock "refresh_cache-2b21e8a1-5f5e-4f1f-bae6-512f1213eb48" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1145.491937] env[62692]: DEBUG nova.compute.manager [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Instance network_info: |[{"id": "144e96aa-d02a-48e4-a91e-73909db5c202", "address": "fa:16:3e:23:7e:07", "network": {"id": "50c1330e-1a5f-42d1-9987-79ddcc0ad7aa", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-813901268-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b1c44eb72b334ee3add29026d312e37b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f92f0b92-d6fb-4d00-8ad5-6b3809ed5493", "external-id": "nsx-vlan-transportzone-48", "segmentation_id": 48, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap144e96aa-d0", "ovs_interfaceid": "144e96aa-d02a-48e4-a91e-73909db5c202", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1145.492464] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:23:7e:07', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f92f0b92-d6fb-4d00-8ad5-6b3809ed5493', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '144e96aa-d02a-48e4-a91e-73909db5c202', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1145.501018] env[62692]: DEBUG oslo.service.loopingcall [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1145.502606] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1145.505038] env[62692]: DEBUG nova.compute.manager [req-4485e584-5bb1-4639-bb9c-639df3a9730c req-ea916740-133a-47a3-81cd-7b3191a584a6 service nova] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Received event network-changed-144e96aa-d02a-48e4-a91e-73909db5c202 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1145.505289] env[62692]: DEBUG nova.compute.manager [req-4485e584-5bb1-4639-bb9c-639df3a9730c req-ea916740-133a-47a3-81cd-7b3191a584a6 service nova] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Refreshing instance network info cache due to event network-changed-144e96aa-d02a-48e4-a91e-73909db5c202. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1145.505565] env[62692]: DEBUG oslo_concurrency.lockutils [req-4485e584-5bb1-4639-bb9c-639df3a9730c req-ea916740-133a-47a3-81cd-7b3191a584a6 service nova] Acquiring lock "refresh_cache-2b21e8a1-5f5e-4f1f-bae6-512f1213eb48" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1145.505754] env[62692]: DEBUG oslo_concurrency.lockutils [req-4485e584-5bb1-4639-bb9c-639df3a9730c req-ea916740-133a-47a3-81cd-7b3191a584a6 service nova] Acquired lock "refresh_cache-2b21e8a1-5f5e-4f1f-bae6-512f1213eb48" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.505981] env[62692]: DEBUG nova.network.neutron [req-4485e584-5bb1-4639-bb9c-639df3a9730c req-ea916740-133a-47a3-81cd-7b3191a584a6 service nova] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Refreshing network info cache for port 144e96aa-d02a-48e4-a91e-73909db5c202 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1145.507260] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1cda6d2d-3dd8-47cc-b8a8-21bce8199fb6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.527690] env[62692]: DEBUG nova.network.neutron [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Updating instance_info_cache with network_info: [{"id": "a668d6b6-4111-4a94-8bc8-09535b29f895", "address": "fa:16:3e:02:7f:2a", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa668d6b6-41", "ovs_interfaceid": "a668d6b6-4111-4a94-8bc8-09535b29f895", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "613360e6-9cd1-4423-9c92-293618e0c0d1", "address": "fa:16:3e:4e:4b:47", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap613360e6-9c", "ovs_interfaceid": "613360e6-9cd1-4423-9c92-293618e0c0d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1145.536692] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1145.536692] env[62692]: value = "task-1141933" [ 1145.536692] env[62692]: _type = "Task" [ 1145.536692] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.545945] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141933, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.583277] env[62692]: DEBUG nova.scheduler.client.report [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1145.657181] env[62692]: DEBUG nova.compute.manager [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1145.679845] env[62692]: DEBUG nova.virt.hardware [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1145.680163] env[62692]: DEBUG nova.virt.hardware [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1145.680336] env[62692]: DEBUG nova.virt.hardware [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1145.680529] env[62692]: DEBUG nova.virt.hardware [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1145.680688] env[62692]: DEBUG nova.virt.hardware [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1145.680892] env[62692]: DEBUG nova.virt.hardware [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1145.681131] env[62692]: DEBUG nova.virt.hardware [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1145.681303] env[62692]: DEBUG nova.virt.hardware [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1145.681487] env[62692]: DEBUG nova.virt.hardware [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1145.681673] env[62692]: DEBUG nova.virt.hardware [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1145.681987] env[62692]: DEBUG nova.virt.hardware [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1145.683009] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e248b1d2-b9c9-4ecb-bc71-3ada875cd7d9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.692028] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd4038e2-3ec0-4ae2-b9a5-4200c9ce1eab {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.750804] env[62692]: DEBUG nova.network.neutron [req-4485e584-5bb1-4639-bb9c-639df3a9730c req-ea916740-133a-47a3-81cd-7b3191a584a6 service nova] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Updated VIF entry in instance network info cache for port 144e96aa-d02a-48e4-a91e-73909db5c202. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1145.751203] env[62692]: DEBUG nova.network.neutron [req-4485e584-5bb1-4639-bb9c-639df3a9730c req-ea916740-133a-47a3-81cd-7b3191a584a6 service nova] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Updating instance_info_cache with network_info: [{"id": "144e96aa-d02a-48e4-a91e-73909db5c202", "address": "fa:16:3e:23:7e:07", "network": {"id": "50c1330e-1a5f-42d1-9987-79ddcc0ad7aa", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-813901268-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b1c44eb72b334ee3add29026d312e37b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f92f0b92-d6fb-4d00-8ad5-6b3809ed5493", "external-id": "nsx-vlan-transportzone-48", "segmentation_id": 48, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap144e96aa-d0", "ovs_interfaceid": "144e96aa-d02a-48e4-a91e-73909db5c202", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1145.906179] env[62692]: DEBUG oslo_concurrency.lockutils [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Releasing lock "refresh_cache-e55a1a5b-3656-48d7-a610-04610b43999a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.030718] env[62692]: DEBUG oslo_concurrency.lockutils [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "refresh_cache-1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.031458] env[62692]: DEBUG oslo_concurrency.lockutils [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1146.031709] env[62692]: DEBUG oslo_concurrency.lockutils [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.032634] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baaa67d2-6fc4-4b90-9dad-1181e5fbe37e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.060769] env[62692]: DEBUG nova.virt.hardware [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1146.061055] env[62692]: DEBUG nova.virt.hardware [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1146.061236] env[62692]: DEBUG nova.virt.hardware [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1146.061439] env[62692]: DEBUG nova.virt.hardware [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1146.061599] env[62692]: DEBUG nova.virt.hardware [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1146.061765] env[62692]: DEBUG nova.virt.hardware [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1146.061987] env[62692]: DEBUG nova.virt.hardware [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1146.062202] env[62692]: DEBUG nova.virt.hardware [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1146.062444] env[62692]: DEBUG nova.virt.hardware [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1146.062589] env[62692]: DEBUG nova.virt.hardware [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1146.062817] env[62692]: DEBUG nova.virt.hardware [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1146.071794] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Reconfiguring VM to attach interface {{(pid=62692) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1146.072174] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141933, 'name': CreateVM_Task, 'duration_secs': 0.425208} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.072402] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b52bfe76-6324-442d-bf54-7ee52ff1712b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.084833] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1146.085931] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1146.086122] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.086478] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1146.087332] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67eddf4c-58c5-4f51-b54c-9aa80b926f53 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.090041] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.462s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1146.090736] env[62692]: DEBUG nova.compute.manager [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1146.095146] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.405s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.095369] env[62692]: DEBUG nova.objects.instance [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Lazy-loading 'resources' on Instance uuid c20139ae-5076-4f6b-b4f0-047332559acf {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1146.096736] env[62692]: DEBUG oslo_vmware.api [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1146.096736] env[62692]: value = "task-1141934" [ 1146.096736] env[62692]: _type = "Task" [ 1146.096736] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.101734] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1146.101734] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]522e3099-21b7-ac3a-6572-1f083d8eb4da" [ 1146.101734] env[62692]: _type = "Task" [ 1146.101734] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.109493] env[62692]: DEBUG oslo_vmware.api [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141934, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.118289] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]522e3099-21b7-ac3a-6572-1f083d8eb4da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.253931] env[62692]: DEBUG oslo_concurrency.lockutils [req-4485e584-5bb1-4639-bb9c-639df3a9730c req-ea916740-133a-47a3-81cd-7b3191a584a6 service nova] Releasing lock "refresh_cache-2b21e8a1-5f5e-4f1f-bae6-512f1213eb48" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.254443] env[62692]: DEBUG nova.compute.manager [req-4485e584-5bb1-4639-bb9c-639df3a9730c req-ea916740-133a-47a3-81cd-7b3191a584a6 service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Received event network-vif-plugged-613360e6-9cd1-4423-9c92-293618e0c0d1 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1146.254847] env[62692]: DEBUG oslo_concurrency.lockutils [req-4485e584-5bb1-4639-bb9c-639df3a9730c req-ea916740-133a-47a3-81cd-7b3191a584a6 service nova] Acquiring lock "1904edb2-fd72-43b6-9219-8043c5abff0c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.255106] env[62692]: DEBUG oslo_concurrency.lockutils [req-4485e584-5bb1-4639-bb9c-639df3a9730c req-ea916740-133a-47a3-81cd-7b3191a584a6 service nova] Lock "1904edb2-fd72-43b6-9219-8043c5abff0c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.255361] env[62692]: DEBUG oslo_concurrency.lockutils [req-4485e584-5bb1-4639-bb9c-639df3a9730c req-ea916740-133a-47a3-81cd-7b3191a584a6 service nova] Lock "1904edb2-fd72-43b6-9219-8043c5abff0c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1146.255704] env[62692]: DEBUG nova.compute.manager [req-4485e584-5bb1-4639-bb9c-639df3a9730c req-ea916740-133a-47a3-81cd-7b3191a584a6 service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] No waiting events found dispatching network-vif-plugged-613360e6-9cd1-4423-9c92-293618e0c0d1 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1146.255941] env[62692]: WARNING nova.compute.manager [req-4485e584-5bb1-4639-bb9c-639df3a9730c req-ea916740-133a-47a3-81cd-7b3191a584a6 service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Received unexpected event network-vif-plugged-613360e6-9cd1-4423-9c92-293618e0c0d1 for instance with vm_state active and task_state None. [ 1146.256211] env[62692]: DEBUG nova.compute.manager [req-4485e584-5bb1-4639-bb9c-639df3a9730c req-ea916740-133a-47a3-81cd-7b3191a584a6 service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Received event network-changed-613360e6-9cd1-4423-9c92-293618e0c0d1 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1146.256489] env[62692]: DEBUG nova.compute.manager [req-4485e584-5bb1-4639-bb9c-639df3a9730c req-ea916740-133a-47a3-81cd-7b3191a584a6 service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Refreshing instance network info cache due to event network-changed-613360e6-9cd1-4423-9c92-293618e0c0d1. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1146.256768] env[62692]: DEBUG oslo_concurrency.lockutils [req-4485e584-5bb1-4639-bb9c-639df3a9730c req-ea916740-133a-47a3-81cd-7b3191a584a6 service nova] Acquiring lock "refresh_cache-1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1146.257022] env[62692]: DEBUG oslo_concurrency.lockutils [req-4485e584-5bb1-4639-bb9c-639df3a9730c req-ea916740-133a-47a3-81cd-7b3191a584a6 service nova] Acquired lock "refresh_cache-1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.257314] env[62692]: DEBUG nova.network.neutron [req-4485e584-5bb1-4639-bb9c-639df3a9730c req-ea916740-133a-47a3-81cd-7b3191a584a6 service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Refreshing network info cache for port 613360e6-9cd1-4423-9c92-293618e0c0d1 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1146.358111] env[62692]: DEBUG nova.network.neutron [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Successfully updated port: 34de5e92-0920-4a35-bb28-51c7d09bf35d {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1146.433804] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed574924-d1cb-4491-b753-ebacba52a081 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.454982] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d98bf1f-da61-4b54-be7f-9896381f110c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.463259] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Updating instance 'e55a1a5b-3656-48d7-a610-04610b43999a' progress to 83 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1146.599668] env[62692]: DEBUG nova.compute.utils [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1146.604159] env[62692]: DEBUG nova.compute.manager [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1146.604359] env[62692]: DEBUG nova.network.neutron [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1146.628630] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]522e3099-21b7-ac3a-6572-1f083d8eb4da, 'name': SearchDatastore_Task, 'duration_secs': 0.015676} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.635770] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.636054] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1146.636302] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1146.636458] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.636685] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1146.637051] env[62692]: DEBUG oslo_vmware.api [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141934, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.640292] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b7cba5cc-4f96-407e-8dd8-285911fedbfe {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.651449] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1146.651449] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1146.654415] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc4a4cc3-b03e-4fee-be75-e0aa51d71003 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.660791] env[62692]: DEBUG nova.policy [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '47c402e4831b4ca6b7db735d109523de', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eb2ee16ec90c48e6babfb1d31ea3de94', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1146.668216] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1146.668216] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5258bbfb-cb93-3fbc-f4e3-718fdaf59d28" [ 1146.668216] env[62692]: _type = "Task" [ 1146.668216] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.673060] env[62692]: DEBUG oslo_vmware.rw_handles [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526094c2-bec6-23ad-7aed-b16318e5bb13/disk-0.vmdk. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1146.674459] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91953616-eb35-4888-8b00-6e5b5f3ac694 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.684825] env[62692]: DEBUG oslo_vmware.rw_handles [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526094c2-bec6-23ad-7aed-b16318e5bb13/disk-0.vmdk is in state: ready. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1146.685025] env[62692]: ERROR oslo_vmware.rw_handles [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526094c2-bec6-23ad-7aed-b16318e5bb13/disk-0.vmdk due to incomplete transfer. [ 1146.688629] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-f0ce4cd6-3224-4102-9af7-607d6ce7051f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.690403] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5258bbfb-cb93-3fbc-f4e3-718fdaf59d28, 'name': SearchDatastore_Task, 'duration_secs': 0.011456} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.694187] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11df8b69-75f4-4542-8683-f8e6fe116674 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.702394] env[62692]: DEBUG oslo_vmware.rw_handles [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526094c2-bec6-23ad-7aed-b16318e5bb13/disk-0.vmdk. {{(pid=62692) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1146.702637] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Uploaded image 59f3345e-2371-4b7f-9944-01f2adf5a997 to the Glance image server {{(pid=62692) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1146.704908] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Destroying the VM {{(pid=62692) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1146.705262] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1146.705262] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52a1261e-b1a2-caa8-d650-ddf26101f208" [ 1146.705262] env[62692]: _type = "Task" [ 1146.705262] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.705464] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-621a276d-09a5-4a00-999b-f4d99c64f3de {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.718216] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52a1261e-b1a2-caa8-d650-ddf26101f208, 'name': SearchDatastore_Task, 'duration_secs': 0.009519} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.721205] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.721486] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48/2b21e8a1-5f5e-4f1f-bae6-512f1213eb48.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1146.721864] env[62692]: DEBUG oslo_vmware.api [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1146.721864] env[62692]: value = "task-1141936" [ 1146.721864] env[62692]: _type = "Task" [ 1146.721864] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.724378] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c701fe83-3ac2-4eb6-84bd-593262111b77 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.737529] env[62692]: DEBUG oslo_vmware.api [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141936, 'name': Destroy_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.741334] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1146.741334] env[62692]: value = "task-1141937" [ 1146.741334] env[62692]: _type = "Task" [ 1146.741334] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.749383] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141937, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.860825] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "refresh_cache-0b7ea87c-252a-48b3-a1cd-c81e91362812" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1146.861038] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquired lock "refresh_cache-0b7ea87c-252a-48b3-a1cd-c81e91362812" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.861215] env[62692]: DEBUG nova.network.neutron [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1146.870910] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67d8e800-3d2a-4dab-9db2-77f054bb748c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.884051] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5321dab0-7515-41e2-a0b2-2b33f25e4ae5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.924186] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c79a390b-5a07-421f-a2d6-d543d7d74092 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.933615] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c509727c-0598-405e-bb39-ea730df983fb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.950394] env[62692]: DEBUG nova.compute.provider_tree [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1146.970787] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1146.971609] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0ffad334-8130-47ec-8e68-7dce686d7ec8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.983600] env[62692]: DEBUG oslo_vmware.api [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1146.983600] env[62692]: value = "task-1141938" [ 1146.983600] env[62692]: _type = "Task" [ 1146.983600] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.997941] env[62692]: DEBUG oslo_vmware.api [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141938, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.003494] env[62692]: DEBUG nova.network.neutron [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Successfully created port: 01bf6621-08ff-4983-8bcb-1d178bda2c60 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1147.069239] env[62692]: DEBUG nova.network.neutron [req-4485e584-5bb1-4639-bb9c-639df3a9730c req-ea916740-133a-47a3-81cd-7b3191a584a6 service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Updated VIF entry in instance network info cache for port 613360e6-9cd1-4423-9c92-293618e0c0d1. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1147.069239] env[62692]: DEBUG nova.network.neutron [req-4485e584-5bb1-4639-bb9c-639df3a9730c req-ea916740-133a-47a3-81cd-7b3191a584a6 service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Updating instance_info_cache with network_info: [{"id": "a668d6b6-4111-4a94-8bc8-09535b29f895", "address": "fa:16:3e:02:7f:2a", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa668d6b6-41", "ovs_interfaceid": "a668d6b6-4111-4a94-8bc8-09535b29f895", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "613360e6-9cd1-4423-9c92-293618e0c0d1", "address": "fa:16:3e:4e:4b:47", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap613360e6-9c", "ovs_interfaceid": "613360e6-9cd1-4423-9c92-293618e0c0d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1147.105053] env[62692]: DEBUG nova.compute.manager [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1147.131584] env[62692]: DEBUG oslo_vmware.api [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141934, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.241846] env[62692]: DEBUG oslo_vmware.api [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141936, 'name': Destroy_Task} progress is 33%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.254080] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141937, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.394927] env[62692]: DEBUG nova.network.neutron [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1147.457530] env[62692]: DEBUG nova.scheduler.client.report [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1147.495760] env[62692]: DEBUG oslo_vmware.api [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141938, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.538648] env[62692]: DEBUG nova.network.neutron [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Updating instance_info_cache with network_info: [{"id": "34de5e92-0920-4a35-bb28-51c7d09bf35d", "address": "fa:16:3e:b7:f9:bc", "network": {"id": "e9c538a3-7723-42cb-85b7-5a1597e49cef", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1878627765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71b00c566cea4b21811ea1e1ef5ba0f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34de5e92-09", "ovs_interfaceid": "34de5e92-0920-4a35-bb28-51c7d09bf35d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1147.572331] env[62692]: DEBUG oslo_concurrency.lockutils [req-4485e584-5bb1-4639-bb9c-639df3a9730c req-ea916740-133a-47a3-81cd-7b3191a584a6 service nova] Releasing lock "refresh_cache-1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1147.616191] env[62692]: DEBUG oslo_vmware.api [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141934, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.740788] env[62692]: DEBUG oslo_vmware.api [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141936, 'name': Destroy_Task, 'duration_secs': 0.642005} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.741089] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Destroyed the VM [ 1147.741342] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Deleting Snapshot of the VM instance {{(pid=62692) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1147.741604] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5c63b4e8-d9f3-486a-aaae-054e09697080 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.752348] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141937, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.563916} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.753513] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48/2b21e8a1-5f5e-4f1f-bae6-512f1213eb48.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1147.753741] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1147.754094] env[62692]: DEBUG oslo_vmware.api [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1147.754094] env[62692]: value = "task-1141939" [ 1147.754094] env[62692]: _type = "Task" [ 1147.754094] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.754355] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3b000246-0fd7-47dd-b6b7-b9ba02d6438e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.766650] env[62692]: DEBUG oslo_vmware.api [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141939, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.768059] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1147.768059] env[62692]: value = "task-1141940" [ 1147.768059] env[62692]: _type = "Task" [ 1147.768059] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.777427] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141940, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.966176] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.871s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.991249] env[62692]: INFO nova.scheduler.client.report [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Deleted allocations for instance c20139ae-5076-4f6b-b4f0-047332559acf [ 1147.998072] env[62692]: DEBUG oslo_vmware.api [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141938, 'name': PowerOnVM_Task, 'duration_secs': 0.647418} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.998601] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1147.998790] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-954a523f-4be4-41c3-aeba-a7cc62b8c007 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Updating instance 'e55a1a5b-3656-48d7-a610-04610b43999a' progress to 100 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1148.042337] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Releasing lock "refresh_cache-0b7ea87c-252a-48b3-a1cd-c81e91362812" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1148.042700] env[62692]: DEBUG nova.compute.manager [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Instance network_info: |[{"id": "34de5e92-0920-4a35-bb28-51c7d09bf35d", "address": "fa:16:3e:b7:f9:bc", "network": {"id": "e9c538a3-7723-42cb-85b7-5a1597e49cef", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1878627765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71b00c566cea4b21811ea1e1ef5ba0f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34de5e92-09", "ovs_interfaceid": "34de5e92-0920-4a35-bb28-51c7d09bf35d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1148.043155] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b7:f9:bc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dd96b39f-bd2e-48d1-85c3-577cf97f08c8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '34de5e92-0920-4a35-bb28-51c7d09bf35d', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1148.050763] env[62692]: DEBUG oslo.service.loopingcall [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1148.051884] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1148.052215] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d9774f3c-2969-4d29-ae1a-b5e999e8f767 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.075234] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1148.075234] env[62692]: value = "task-1141941" [ 1148.075234] env[62692]: _type = "Task" [ 1148.075234] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.086284] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141941, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.117344] env[62692]: DEBUG oslo_vmware.api [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141934, 'name': ReconfigVM_Task, 'duration_secs': 1.621721} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.117870] env[62692]: DEBUG oslo_concurrency.lockutils [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1148.118098] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Reconfigured VM to attach interface {{(pid=62692) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1148.125142] env[62692]: DEBUG nova.compute.manager [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1148.151013] env[62692]: DEBUG nova.virt.hardware [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1148.151291] env[62692]: DEBUG nova.virt.hardware [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1148.151459] env[62692]: DEBUG nova.virt.hardware [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1148.151696] env[62692]: DEBUG nova.virt.hardware [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1148.151884] env[62692]: DEBUG nova.virt.hardware [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1148.152122] env[62692]: DEBUG nova.virt.hardware [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1148.152373] env[62692]: DEBUG nova.virt.hardware [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1148.152542] env[62692]: DEBUG nova.virt.hardware [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1148.152735] env[62692]: DEBUG nova.virt.hardware [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1148.153343] env[62692]: DEBUG nova.virt.hardware [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1148.153343] env[62692]: DEBUG nova.virt.hardware [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1148.154371] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9179f30-a306-4114-84f6-262de91d1b93 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.163164] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffde8b5e-6abe-49ea-be69-838460be0af1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.268337] env[62692]: DEBUG oslo_vmware.api [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141939, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.276704] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141940, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088565} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.276959] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1148.277763] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea60b625-4d91-4d5d-896a-dcebcb198df0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.300197] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48/2b21e8a1-5f5e-4f1f-bae6-512f1213eb48.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1148.300444] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-35a1eec0-5c57-4191-86ed-05d1d71ee364 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.321853] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1148.321853] env[62692]: value = "task-1141942" [ 1148.321853] env[62692]: _type = "Task" [ 1148.321853] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.329888] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141942, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.477470] env[62692]: DEBUG nova.compute.manager [req-48484b83-9e6a-46ec-9f0e-f01e685d3f29 req-33b759c6-57fc-4ace-99e9-dbfcd4bfb93b service nova] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Received event network-vif-plugged-34de5e92-0920-4a35-bb28-51c7d09bf35d {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1148.477634] env[62692]: DEBUG oslo_concurrency.lockutils [req-48484b83-9e6a-46ec-9f0e-f01e685d3f29 req-33b759c6-57fc-4ace-99e9-dbfcd4bfb93b service nova] Acquiring lock "0b7ea87c-252a-48b3-a1cd-c81e91362812-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1148.477789] env[62692]: DEBUG oslo_concurrency.lockutils [req-48484b83-9e6a-46ec-9f0e-f01e685d3f29 req-33b759c6-57fc-4ace-99e9-dbfcd4bfb93b service nova] Lock "0b7ea87c-252a-48b3-a1cd-c81e91362812-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1148.477891] env[62692]: DEBUG oslo_concurrency.lockutils [req-48484b83-9e6a-46ec-9f0e-f01e685d3f29 req-33b759c6-57fc-4ace-99e9-dbfcd4bfb93b service nova] Lock "0b7ea87c-252a-48b3-a1cd-c81e91362812-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1148.478161] env[62692]: DEBUG nova.compute.manager [req-48484b83-9e6a-46ec-9f0e-f01e685d3f29 req-33b759c6-57fc-4ace-99e9-dbfcd4bfb93b service nova] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] No waiting events found dispatching network-vif-plugged-34de5e92-0920-4a35-bb28-51c7d09bf35d {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1148.478384] env[62692]: WARNING nova.compute.manager [req-48484b83-9e6a-46ec-9f0e-f01e685d3f29 req-33b759c6-57fc-4ace-99e9-dbfcd4bfb93b service nova] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Received unexpected event network-vif-plugged-34de5e92-0920-4a35-bb28-51c7d09bf35d for instance with vm_state building and task_state spawning. [ 1148.478550] env[62692]: DEBUG nova.compute.manager [req-48484b83-9e6a-46ec-9f0e-f01e685d3f29 req-33b759c6-57fc-4ace-99e9-dbfcd4bfb93b service nova] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Received event network-changed-34de5e92-0920-4a35-bb28-51c7d09bf35d {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1148.479524] env[62692]: DEBUG nova.compute.manager [req-48484b83-9e6a-46ec-9f0e-f01e685d3f29 req-33b759c6-57fc-4ace-99e9-dbfcd4bfb93b service nova] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Refreshing instance network info cache due to event network-changed-34de5e92-0920-4a35-bb28-51c7d09bf35d. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1148.479524] env[62692]: DEBUG oslo_concurrency.lockutils [req-48484b83-9e6a-46ec-9f0e-f01e685d3f29 req-33b759c6-57fc-4ace-99e9-dbfcd4bfb93b service nova] Acquiring lock "refresh_cache-0b7ea87c-252a-48b3-a1cd-c81e91362812" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1148.479524] env[62692]: DEBUG oslo_concurrency.lockutils [req-48484b83-9e6a-46ec-9f0e-f01e685d3f29 req-33b759c6-57fc-4ace-99e9-dbfcd4bfb93b service nova] Acquired lock "refresh_cache-0b7ea87c-252a-48b3-a1cd-c81e91362812" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1148.479524] env[62692]: DEBUG nova.network.neutron [req-48484b83-9e6a-46ec-9f0e-f01e685d3f29 req-33b759c6-57fc-4ace-99e9-dbfcd4bfb93b service nova] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Refreshing network info cache for port 34de5e92-0920-4a35-bb28-51c7d09bf35d {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1148.502575] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e0a25288-711a-4942-99f5-c223aff0721a tempest-ServerPasswordTestJSON-1547161886 tempest-ServerPasswordTestJSON-1547161886-project-member] Lock "c20139ae-5076-4f6b-b4f0-047332559acf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.733s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1148.562766] env[62692]: DEBUG nova.network.neutron [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Port a3885c51-6f7b-41e8-8948-41fac442a288 binding to destination host cpu-1 is already ACTIVE {{(pid=62692) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1148.563091] env[62692]: DEBUG oslo_concurrency.lockutils [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "refresh_cache-cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1148.563254] env[62692]: DEBUG oslo_concurrency.lockutils [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired lock "refresh_cache-cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1148.563439] env[62692]: DEBUG nova.network.neutron [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1148.586925] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141941, 'name': CreateVM_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.623116] env[62692]: DEBUG oslo_concurrency.lockutils [None req-40c1282e-7ed1-466c-b54a-7c9c0e5038b6 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "interface-1904edb2-fd72-43b6-9219-8043c5abff0c-613360e6-9cd1-4423-9c92-293618e0c0d1" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.722s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1148.770414] env[62692]: DEBUG oslo_vmware.api [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141939, 'name': RemoveSnapshot_Task} progress is 16%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.830874] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141942, 'name': ReconfigVM_Task, 'duration_secs': 0.447413} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.831434] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48/2b21e8a1-5f5e-4f1f-bae6-512f1213eb48.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1148.832142] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1165f257-8552-4f01-81bf-7d6f570d17ef {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.840144] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1148.840144] env[62692]: value = "task-1141944" [ 1148.840144] env[62692]: _type = "Task" [ 1148.840144] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.848644] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141944, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.035467] env[62692]: DEBUG nova.network.neutron [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Successfully updated port: 01bf6621-08ff-4983-8bcb-1d178bda2c60 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1149.088806] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141941, 'name': CreateVM_Task, 'duration_secs': 0.635917} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.089236] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1149.089855] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1149.090034] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1149.090385] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1149.090951] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09fcae2b-69ca-457c-ab9e-a946b77ef5e4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.096200] env[62692]: DEBUG oslo_vmware.api [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1149.096200] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52469fb2-eca7-e99d-e48b-d19aaeb26e73" [ 1149.096200] env[62692]: _type = "Task" [ 1149.096200] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.107020] env[62692]: DEBUG oslo_vmware.api [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52469fb2-eca7-e99d-e48b-d19aaeb26e73, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.202896] env[62692]: DEBUG nova.network.neutron [req-48484b83-9e6a-46ec-9f0e-f01e685d3f29 req-33b759c6-57fc-4ace-99e9-dbfcd4bfb93b service nova] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Updated VIF entry in instance network info cache for port 34de5e92-0920-4a35-bb28-51c7d09bf35d. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1149.203360] env[62692]: DEBUG nova.network.neutron [req-48484b83-9e6a-46ec-9f0e-f01e685d3f29 req-33b759c6-57fc-4ace-99e9-dbfcd4bfb93b service nova] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Updating instance_info_cache with network_info: [{"id": "34de5e92-0920-4a35-bb28-51c7d09bf35d", "address": "fa:16:3e:b7:f9:bc", "network": {"id": "e9c538a3-7723-42cb-85b7-5a1597e49cef", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1878627765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71b00c566cea4b21811ea1e1ef5ba0f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34de5e92-09", "ovs_interfaceid": "34de5e92-0920-4a35-bb28-51c7d09bf35d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1149.268364] env[62692]: DEBUG oslo_vmware.api [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141939, 'name': RemoveSnapshot_Task, 'duration_secs': 1.162104} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.268639] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Deleted Snapshot of the VM instance {{(pid=62692) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1149.268880] env[62692]: INFO nova.compute.manager [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Took 16.01 seconds to snapshot the instance on the hypervisor. [ 1149.349377] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141944, 'name': Rename_Task, 'duration_secs': 0.247469} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.349672] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1149.349930] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a22139c9-5015-4f64-83fe-566a6cf794a6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.357171] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1149.357171] env[62692]: value = "task-1141945" [ 1149.357171] env[62692]: _type = "Task" [ 1149.357171] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.365665] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141945, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.487895] env[62692]: DEBUG nova.network.neutron [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Updating instance_info_cache with network_info: [{"id": "a3885c51-6f7b-41e8-8948-41fac442a288", "address": "fa:16:3e:e4:9b:6f", "network": {"id": "f70727d6-a9aa-44b0-a244-2533b3f96571", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-284105837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1037e5d436bd429391159b87a712f6e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3885c51-6f", "ovs_interfaceid": "a3885c51-6f7b-41e8-8948-41fac442a288", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1149.544450] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "refresh_cache-45eccffc-f059-4bef-97da-dacb16f20f88" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1149.544450] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquired lock "refresh_cache-45eccffc-f059-4bef-97da-dacb16f20f88" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1149.544450] env[62692]: DEBUG nova.network.neutron [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1149.610472] env[62692]: DEBUG oslo_vmware.api [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52469fb2-eca7-e99d-e48b-d19aaeb26e73, 'name': SearchDatastore_Task, 'duration_secs': 0.020196} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.610793] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1149.611287] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1149.611592] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1149.611810] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1149.612169] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1149.612470] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3a730338-8184-41fa-80db-5cb048ce1487 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.624726] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1149.624940] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1149.625724] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b3dbdb7-017d-4d79-88a5-e170ba22891f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.633041] env[62692]: DEBUG oslo_vmware.api [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1149.633041] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]529c0d02-12fe-7094-9412-824c765fe97b" [ 1149.633041] env[62692]: _type = "Task" [ 1149.633041] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.641584] env[62692]: DEBUG oslo_vmware.api [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]529c0d02-12fe-7094-9412-824c765fe97b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.706442] env[62692]: DEBUG oslo_concurrency.lockutils [req-48484b83-9e6a-46ec-9f0e-f01e685d3f29 req-33b759c6-57fc-4ace-99e9-dbfcd4bfb93b service nova] Releasing lock "refresh_cache-0b7ea87c-252a-48b3-a1cd-c81e91362812" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1149.845160] env[62692]: DEBUG nova.compute.manager [None req-750a41c7-4970-419b-aa31-f933b61e4c2c tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Found 2 images (rotation: 2) {{(pid=62692) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1149.868392] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141945, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.993361] env[62692]: DEBUG oslo_concurrency.lockutils [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Releasing lock "refresh_cache-cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1150.079094] env[62692]: DEBUG nova.network.neutron [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1150.145175] env[62692]: DEBUG oslo_vmware.api [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]529c0d02-12fe-7094-9412-824c765fe97b, 'name': SearchDatastore_Task, 'duration_secs': 0.028313} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.146042] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3154c65e-99c5-4227-a81f-68d41f9d7d4d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.152606] env[62692]: DEBUG oslo_vmware.api [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1150.152606] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52f0a373-b20f-5e01-dce2-d43a2be3b1f2" [ 1150.152606] env[62692]: _type = "Task" [ 1150.152606] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.160940] env[62692]: DEBUG oslo_vmware.api [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52f0a373-b20f-5e01-dce2-d43a2be3b1f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.188434] env[62692]: DEBUG oslo_concurrency.lockutils [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "interface-1904edb2-fd72-43b6-9219-8043c5abff0c-613360e6-9cd1-4423-9c92-293618e0c0d1" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1150.188434] env[62692]: DEBUG oslo_concurrency.lockutils [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "interface-1904edb2-fd72-43b6-9219-8043c5abff0c-613360e6-9cd1-4423-9c92-293618e0c0d1" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1150.288468] env[62692]: DEBUG nova.network.neutron [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Updating instance_info_cache with network_info: [{"id": "01bf6621-08ff-4983-8bcb-1d178bda2c60", "address": "fa:16:3e:15:0e:b2", "network": {"id": "26fe2f12-b216-4fa5-a022-24c5ab315416", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-838584488-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb2ee16ec90c48e6babfb1d31ea3de94", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01bf6621-08", "ovs_interfaceid": "01bf6621-08ff-4983-8bcb-1d178bda2c60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1150.374341] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141945, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.494973] env[62692]: DEBUG nova.compute.manager [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62692) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1150.495175] env[62692]: DEBUG oslo_concurrency.lockutils [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1150.495435] env[62692]: DEBUG oslo_concurrency.lockutils [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1150.663390] env[62692]: DEBUG oslo_vmware.api [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52f0a373-b20f-5e01-dce2-d43a2be3b1f2, 'name': SearchDatastore_Task, 'duration_secs': 0.024747} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.663736] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1150.664152] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 0b7ea87c-252a-48b3-a1cd-c81e91362812/0b7ea87c-252a-48b3-a1cd-c81e91362812.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1150.664523] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bdd1761d-da6c-4266-87a5-58da2c2eb4d1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.674473] env[62692]: DEBUG oslo_vmware.api [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1150.674473] env[62692]: value = "task-1141946" [ 1150.674473] env[62692]: _type = "Task" [ 1150.674473] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.684390] env[62692]: DEBUG oslo_vmware.api [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141946, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.692198] env[62692]: DEBUG oslo_concurrency.lockutils [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1150.692467] env[62692]: DEBUG oslo_concurrency.lockutils [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1150.693348] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c917880-41b4-42fc-bfdb-f6335a288bdc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.714752] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ac3cf47-1ec3-4965-8c19-734db0ceb533 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.742092] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Reconfiguring VM to detach interface {{(pid=62692) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1150.742577] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-05c07be6-96ea-4787-bb90-f4a4c81159c4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.764878] env[62692]: DEBUG oslo_vmware.api [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1150.764878] env[62692]: value = "task-1141948" [ 1150.764878] env[62692]: _type = "Task" [ 1150.764878] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.775915] env[62692]: DEBUG oslo_vmware.api [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141948, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.791150] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Releasing lock "refresh_cache-45eccffc-f059-4bef-97da-dacb16f20f88" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1150.791629] env[62692]: DEBUG nova.compute.manager [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Instance network_info: |[{"id": "01bf6621-08ff-4983-8bcb-1d178bda2c60", "address": "fa:16:3e:15:0e:b2", "network": {"id": "26fe2f12-b216-4fa5-a022-24c5ab315416", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-838584488-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb2ee16ec90c48e6babfb1d31ea3de94", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01bf6621-08", "ovs_interfaceid": "01bf6621-08ff-4983-8bcb-1d178bda2c60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1150.792172] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:15:0e:b2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0636c3f6-fcb7-4954-ab07-c5cd0dee37b0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '01bf6621-08ff-4983-8bcb-1d178bda2c60', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1150.801471] env[62692]: DEBUG oslo.service.loopingcall [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1150.802014] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1150.802160] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fa952e48-7660-4908-bac4-0275118da45d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.821661] env[62692]: DEBUG nova.compute.manager [req-382ed533-feaa-480a-8d55-edfc1ed20000 req-27ae94b4-acf5-49d7-99e5-323be5432fc6 service nova] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Received event network-vif-plugged-01bf6621-08ff-4983-8bcb-1d178bda2c60 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1150.821919] env[62692]: DEBUG oslo_concurrency.lockutils [req-382ed533-feaa-480a-8d55-edfc1ed20000 req-27ae94b4-acf5-49d7-99e5-323be5432fc6 service nova] Acquiring lock "45eccffc-f059-4bef-97da-dacb16f20f88-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1150.822159] env[62692]: DEBUG oslo_concurrency.lockutils [req-382ed533-feaa-480a-8d55-edfc1ed20000 req-27ae94b4-acf5-49d7-99e5-323be5432fc6 service nova] Lock "45eccffc-f059-4bef-97da-dacb16f20f88-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1150.822337] env[62692]: DEBUG oslo_concurrency.lockutils [req-382ed533-feaa-480a-8d55-edfc1ed20000 req-27ae94b4-acf5-49d7-99e5-323be5432fc6 service nova] Lock "45eccffc-f059-4bef-97da-dacb16f20f88-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1150.822511] env[62692]: DEBUG nova.compute.manager [req-382ed533-feaa-480a-8d55-edfc1ed20000 req-27ae94b4-acf5-49d7-99e5-323be5432fc6 service nova] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] No waiting events found dispatching network-vif-plugged-01bf6621-08ff-4983-8bcb-1d178bda2c60 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1150.822677] env[62692]: WARNING nova.compute.manager [req-382ed533-feaa-480a-8d55-edfc1ed20000 req-27ae94b4-acf5-49d7-99e5-323be5432fc6 service nova] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Received unexpected event network-vif-plugged-01bf6621-08ff-4983-8bcb-1d178bda2c60 for instance with vm_state building and task_state spawning. [ 1150.822882] env[62692]: DEBUG nova.compute.manager [req-382ed533-feaa-480a-8d55-edfc1ed20000 req-27ae94b4-acf5-49d7-99e5-323be5432fc6 service nova] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Received event network-changed-01bf6621-08ff-4983-8bcb-1d178bda2c60 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1150.823067] env[62692]: DEBUG nova.compute.manager [req-382ed533-feaa-480a-8d55-edfc1ed20000 req-27ae94b4-acf5-49d7-99e5-323be5432fc6 service nova] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Refreshing instance network info cache due to event network-changed-01bf6621-08ff-4983-8bcb-1d178bda2c60. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1150.823369] env[62692]: DEBUG oslo_concurrency.lockutils [req-382ed533-feaa-480a-8d55-edfc1ed20000 req-27ae94b4-acf5-49d7-99e5-323be5432fc6 service nova] Acquiring lock "refresh_cache-45eccffc-f059-4bef-97da-dacb16f20f88" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1150.823580] env[62692]: DEBUG oslo_concurrency.lockutils [req-382ed533-feaa-480a-8d55-edfc1ed20000 req-27ae94b4-acf5-49d7-99e5-323be5432fc6 service nova] Acquired lock "refresh_cache-45eccffc-f059-4bef-97da-dacb16f20f88" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1150.823720] env[62692]: DEBUG nova.network.neutron [req-382ed533-feaa-480a-8d55-edfc1ed20000 req-27ae94b4-acf5-49d7-99e5-323be5432fc6 service nova] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Refreshing network info cache for port 01bf6621-08ff-4983-8bcb-1d178bda2c60 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1150.826610] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1150.826610] env[62692]: value = "task-1141949" [ 1150.826610] env[62692]: _type = "Task" [ 1150.826610] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.837259] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141949, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.861845] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "e55a1a5b-3656-48d7-a610-04610b43999a" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1150.862218] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "e55a1a5b-3656-48d7-a610-04610b43999a" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1150.862458] env[62692]: DEBUG nova.compute.manager [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Going to confirm migration 6 {{(pid=62692) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1150.874255] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141945, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.999725] env[62692]: DEBUG nova.objects.instance [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lazy-loading 'migration_context' on Instance uuid cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1151.012332] env[62692]: DEBUG nova.compute.manager [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1151.013509] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-965dfb81-5dfd-4a2f-b751-6733edd2f980 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.189101] env[62692]: DEBUG oslo_vmware.api [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141946, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.276402] env[62692]: DEBUG oslo_vmware.api [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141948, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.341386] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141949, 'name': CreateVM_Task, 'duration_secs': 0.481283} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.341386] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1151.342133] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1151.342308] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1151.342651] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1151.342940] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-856d3b25-0e56-41b2-99a9-462aafb1ada1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.348323] env[62692]: DEBUG oslo_vmware.api [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1151.348323] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52b86dc3-906c-2362-5400-48c50acacf70" [ 1151.348323] env[62692]: _type = "Task" [ 1151.348323] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.356996] env[62692]: DEBUG oslo_vmware.api [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52b86dc3-906c-2362-5400-48c50acacf70, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.374148] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141945, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.439449] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "refresh_cache-e55a1a5b-3656-48d7-a610-04610b43999a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1151.439722] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquired lock "refresh_cache-e55a1a5b-3656-48d7-a610-04610b43999a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1151.439951] env[62692]: DEBUG nova.network.neutron [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1151.440245] env[62692]: DEBUG nova.objects.instance [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lazy-loading 'info_cache' on Instance uuid e55a1a5b-3656-48d7-a610-04610b43999a {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1151.525648] env[62692]: INFO nova.compute.manager [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] instance snapshotting [ 1151.526454] env[62692]: DEBUG nova.objects.instance [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lazy-loading 'flavor' on Instance uuid 90932f05-fad3-4c6d-87ae-ab059351b0be {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1151.552880] env[62692]: DEBUG nova.network.neutron [req-382ed533-feaa-480a-8d55-edfc1ed20000 req-27ae94b4-acf5-49d7-99e5-323be5432fc6 service nova] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Updated VIF entry in instance network info cache for port 01bf6621-08ff-4983-8bcb-1d178bda2c60. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1151.553133] env[62692]: DEBUG nova.network.neutron [req-382ed533-feaa-480a-8d55-edfc1ed20000 req-27ae94b4-acf5-49d7-99e5-323be5432fc6 service nova] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Updating instance_info_cache with network_info: [{"id": "01bf6621-08ff-4983-8bcb-1d178bda2c60", "address": "fa:16:3e:15:0e:b2", "network": {"id": "26fe2f12-b216-4fa5-a022-24c5ab315416", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-838584488-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb2ee16ec90c48e6babfb1d31ea3de94", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01bf6621-08", "ovs_interfaceid": "01bf6621-08ff-4983-8bcb-1d178bda2c60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1151.687310] env[62692]: DEBUG oslo_vmware.api [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141946, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.548981} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.687596] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 0b7ea87c-252a-48b3-a1cd-c81e91362812/0b7ea87c-252a-48b3-a1cd-c81e91362812.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1151.687821] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1151.688057] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-463eb8cc-0b04-4d0c-9893-6f9bb6a683ad {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.691214] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aeaa86d-f422-48e9-a2be-f9ca5e1b915c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.702260] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6b190ba-7148-4157-9094-f8564f40c81f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.705899] env[62692]: DEBUG oslo_vmware.api [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1151.705899] env[62692]: value = "task-1141950" [ 1151.705899] env[62692]: _type = "Task" [ 1151.705899] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.737256] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-429232be-263d-44b0-be98-d7fa76346ac1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.744028] env[62692]: DEBUG oslo_vmware.api [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141950, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.750870] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f44af1-c61f-42ff-9bf0-354fe7d45654 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.768163] env[62692]: DEBUG nova.compute.provider_tree [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1151.778174] env[62692]: DEBUG oslo_vmware.api [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141948, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.859788] env[62692]: DEBUG oslo_vmware.api [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52b86dc3-906c-2362-5400-48c50acacf70, 'name': SearchDatastore_Task, 'duration_secs': 0.021741} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.860138] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1151.860385] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1151.860623] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1151.860775] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1151.861052] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1151.861271] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9971ceb2-2f05-48ba-950e-7ca1bb323c02 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.871895] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141945, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.875081] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1151.875278] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1151.875976] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c84b0abf-8c5a-4eb1-a4fd-e521d62f787e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.881076] env[62692]: DEBUG oslo_vmware.api [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1151.881076] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5286d86f-cccb-127d-fae6-65128677586e" [ 1151.881076] env[62692]: _type = "Task" [ 1151.881076] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.889758] env[62692]: DEBUG oslo_vmware.api [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5286d86f-cccb-127d-fae6-65128677586e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.032502] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52a49def-8d90-48d8-802c-d526f4eb4bd1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.054129] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b1308a3-11de-439f-86d6-ff938410420f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.057849] env[62692]: DEBUG oslo_concurrency.lockutils [req-382ed533-feaa-480a-8d55-edfc1ed20000 req-27ae94b4-acf5-49d7-99e5-323be5432fc6 service nova] Releasing lock "refresh_cache-45eccffc-f059-4bef-97da-dacb16f20f88" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1152.215971] env[62692]: DEBUG oslo_vmware.api [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141950, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071397} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.216295] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1152.217081] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3f96cb9-46dc-43a9-913d-cb8e2af4fb7c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.239267] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] 0b7ea87c-252a-48b3-a1cd-c81e91362812/0b7ea87c-252a-48b3-a1cd-c81e91362812.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1152.239700] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3cbb1aa-e9be-4495-879e-cf291b487f0d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.260972] env[62692]: DEBUG oslo_vmware.api [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1152.260972] env[62692]: value = "task-1141951" [ 1152.260972] env[62692]: _type = "Task" [ 1152.260972] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.270342] env[62692]: DEBUG oslo_vmware.api [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141951, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.274233] env[62692]: DEBUG nova.scheduler.client.report [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1152.283968] env[62692]: DEBUG oslo_vmware.api [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141948, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.378226] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141945, 'name': PowerOnVM_Task} progress is 81%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.394130] env[62692]: DEBUG oslo_vmware.api [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5286d86f-cccb-127d-fae6-65128677586e, 'name': SearchDatastore_Task, 'duration_secs': 0.024365} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.395322] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e907054d-88d1-4b42-9865-aaf74c217ac8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.402989] env[62692]: DEBUG oslo_vmware.api [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1152.402989] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5271ad76-982e-419a-772f-595ea864e0cd" [ 1152.402989] env[62692]: _type = "Task" [ 1152.402989] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.414512] env[62692]: DEBUG oslo_vmware.api [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5271ad76-982e-419a-772f-595ea864e0cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.565582] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Creating Snapshot of the VM instance {{(pid=62692) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1152.568448] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8bfa2493-19d5-48e7-93f7-d44145d53c75 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.578541] env[62692]: DEBUG oslo_vmware.api [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1152.578541] env[62692]: value = "task-1141952" [ 1152.578541] env[62692]: _type = "Task" [ 1152.578541] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.587599] env[62692]: DEBUG oslo_vmware.api [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141952, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.702265] env[62692]: DEBUG nova.network.neutron [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Updating instance_info_cache with network_info: [{"id": "5e50f271-31a9-4755-a612-6b971179167a", "address": "fa:16:3e:75:3c:39", "network": {"id": "dd260020-2d74-46ba-aa78-af29bdefa176", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-617179655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "589895ae20394e2fa525707bf7d09f2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e50f271-31", "ovs_interfaceid": "5e50f271-31a9-4755-a612-6b971179167a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1152.772299] env[62692]: DEBUG oslo_vmware.api [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141951, 'name': ReconfigVM_Task, 'duration_secs': 0.504857} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.776508] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Reconfigured VM instance instance-0000006e to attach disk [datastore2] 0b7ea87c-252a-48b3-a1cd-c81e91362812/0b7ea87c-252a-48b3-a1cd-c81e91362812.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1152.776508] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9275e47f-f673-4160-a40e-f54a275fbd89 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.788646] env[62692]: DEBUG oslo_vmware.api [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141948, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.804161] env[62692]: DEBUG oslo_vmware.api [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1152.804161] env[62692]: value = "task-1141953" [ 1152.804161] env[62692]: _type = "Task" [ 1152.804161] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.815823] env[62692]: DEBUG oslo_vmware.api [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141953, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.876915] env[62692]: DEBUG oslo_vmware.api [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141945, 'name': PowerOnVM_Task, 'duration_secs': 3.370927} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.876915] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1152.878839] env[62692]: INFO nova.compute.manager [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Took 9.94 seconds to spawn the instance on the hypervisor. [ 1152.878839] env[62692]: DEBUG nova.compute.manager [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1152.878839] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3248eb76-893b-4376-b108-8d715e0242bf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.916086] env[62692]: DEBUG oslo_vmware.api [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5271ad76-982e-419a-772f-595ea864e0cd, 'name': SearchDatastore_Task, 'duration_secs': 0.047263} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.916434] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1152.916751] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 45eccffc-f059-4bef-97da-dacb16f20f88/45eccffc-f059-4bef-97da-dacb16f20f88.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1152.917624] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a85a187a-b8c2-4b4d-9c1e-038a67955df6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.926383] env[62692]: DEBUG oslo_vmware.api [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1152.926383] env[62692]: value = "task-1141954" [ 1152.926383] env[62692]: _type = "Task" [ 1152.926383] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.937423] env[62692]: DEBUG oslo_vmware.api [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141954, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.091849] env[62692]: DEBUG oslo_vmware.api [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141952, 'name': CreateSnapshot_Task, 'duration_secs': 0.509105} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.092159] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Created Snapshot of the VM instance {{(pid=62692) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1153.092972] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0883c5e-ffae-4a05-83f2-5427902e4816 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.178388] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1153.178695] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1153.178890] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Starting heal instance info cache {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1153.179087] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Rebuilding the list of instances to heal {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1153.206558] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Releasing lock "refresh_cache-e55a1a5b-3656-48d7-a610-04610b43999a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1153.206916] env[62692]: DEBUG nova.objects.instance [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lazy-loading 'migration_context' on Instance uuid e55a1a5b-3656-48d7-a610-04610b43999a {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1153.289293] env[62692]: DEBUG oslo_vmware.api [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141948, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.293037] env[62692]: DEBUG oslo_concurrency.lockutils [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.795s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.321792] env[62692]: DEBUG oslo_vmware.api [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141953, 'name': Rename_Task, 'duration_secs': 0.162363} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.324178] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1153.324178] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e0c79144-66fc-48ed-aaa7-b5f2cddde848 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.334158] env[62692]: DEBUG oslo_vmware.api [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1153.334158] env[62692]: value = "task-1141955" [ 1153.334158] env[62692]: _type = "Task" [ 1153.334158] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.343552] env[62692]: DEBUG oslo_vmware.api [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141955, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.397995] env[62692]: INFO nova.compute.manager [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Took 15.84 seconds to build instance. [ 1153.438551] env[62692]: DEBUG oslo_vmware.api [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141954, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.611687] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Creating linked-clone VM from snapshot {{(pid=62692) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1153.612129] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0b2a955d-cea7-42d2-9b49-3cd9fdac6fc0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.622755] env[62692]: DEBUG oslo_vmware.api [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1153.622755] env[62692]: value = "task-1141956" [ 1153.622755] env[62692]: _type = "Task" [ 1153.622755] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.632404] env[62692]: DEBUG oslo_vmware.api [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141956, 'name': CloneVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.688495] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Skipping network cache update for instance because it is being deleted. {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9943}} [ 1153.688709] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Skipping network cache update for instance because it is Building. {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 1153.688863] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Skipping network cache update for instance because it is Building. {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 1153.713916] env[62692]: DEBUG nova.objects.base [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62692) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1153.713916] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f7dc5ae-75d5-4405-9f93-b1c533d35fd9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.733510] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "refresh_cache-e2383a6a-3581-40fc-a0eb-6981acdbf54a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1153.733510] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquired lock "refresh_cache-e2383a6a-3581-40fc-a0eb-6981acdbf54a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.733510] env[62692]: DEBUG nova.network.neutron [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Forcefully refreshing network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1153.733740] env[62692]: DEBUG nova.objects.instance [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lazy-loading 'info_cache' on Instance uuid e2383a6a-3581-40fc-a0eb-6981acdbf54a {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1153.738150] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77258c3a-2148-413c-8e0d-13a881d217c3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.748487] env[62692]: DEBUG oslo_vmware.api [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1153.748487] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5216026f-4edf-6203-6617-845fef79fff2" [ 1153.748487] env[62692]: _type = "Task" [ 1153.748487] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.751927] env[62692]: DEBUG oslo_vmware.api [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5216026f-4edf-6203-6617-845fef79fff2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.784561] env[62692]: DEBUG oslo_vmware.api [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141948, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.847975] env[62692]: DEBUG oslo_vmware.api [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141955, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.900265] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6e6f0e8e-99b3-41e8-a135-8fcc7519dc51 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Lock "2b21e8a1-5f5e-4f1f-bae6-512f1213eb48" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.352s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.938481] env[62692]: DEBUG oslo_vmware.api [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141954, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.548386} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.938867] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 45eccffc-f059-4bef-97da-dacb16f20f88/45eccffc-f059-4bef-97da-dacb16f20f88.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1153.939016] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1153.939274] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e35fd70d-4da0-4561-976f-c99f69a7903f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.947857] env[62692]: DEBUG oslo_vmware.api [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1153.947857] env[62692]: value = "task-1141957" [ 1153.947857] env[62692]: _type = "Task" [ 1153.947857] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.957764] env[62692]: DEBUG oslo_vmware.api [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141957, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.133571] env[62692]: DEBUG oslo_vmware.api [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141956, 'name': CloneVM_Task} progress is 94%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.260287] env[62692]: DEBUG oslo_vmware.api [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5216026f-4edf-6203-6617-845fef79fff2, 'name': SearchDatastore_Task, 'duration_secs': 0.016135} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.261576] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1154.261854] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1154.286244] env[62692]: DEBUG oslo_vmware.api [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141948, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.346370] env[62692]: DEBUG oslo_vmware.api [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141955, 'name': PowerOnVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.458873] env[62692]: DEBUG oslo_vmware.api [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141957, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085031} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.459177] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1154.460090] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6878a28-1d23-4157-aa70-83c9088eb77c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.484270] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] 45eccffc-f059-4bef-97da-dacb16f20f88/45eccffc-f059-4bef-97da-dacb16f20f88.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1154.484501] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e29143b-31b5-477c-9cff-3d51a352293a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.506532] env[62692]: DEBUG oslo_vmware.api [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1154.506532] env[62692]: value = "task-1141958" [ 1154.506532] env[62692]: _type = "Task" [ 1154.506532] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.515378] env[62692]: DEBUG oslo_vmware.api [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141958, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.635527] env[62692]: DEBUG oslo_vmware.api [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141956, 'name': CloneVM_Task} progress is 94%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.789136] env[62692]: DEBUG oslo_vmware.api [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141948, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.850351] env[62692]: DEBUG oslo_vmware.api [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1141955, 'name': PowerOnVM_Task, 'duration_secs': 1.512407} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.850729] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1154.850943] env[62692]: INFO nova.compute.manager [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Took 9.19 seconds to spawn the instance on the hypervisor. [ 1154.851151] env[62692]: DEBUG nova.compute.manager [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1154.852348] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce3806b4-4ea8-4919-bd69-1669714856e3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.857642] env[62692]: INFO nova.compute.manager [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Swapping old allocation on dict_keys(['470758a8-5efa-4875-aad5-f512a727752f']) held by migration 7933eebe-a34a-46f0-8c3c-80776a5c4adc for instance [ 1154.893378] env[62692]: DEBUG nova.scheduler.client.report [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Overwriting current allocation {'allocations': {'470758a8-5efa-4875-aad5-f512a727752f': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 147}}, 'project_id': '1037e5d436bd429391159b87a712f6e1', 'user_id': '6e9cd6d33e744da4a2f678226b684845', 'consumer_generation': 1} on consumer cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43 {{(pid=62692) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1155.009042] env[62692]: DEBUG oslo_concurrency.lockutils [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "973ac4ab-5f82-4108-8076-9a370ca9d914" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.009311] env[62692]: DEBUG oslo_concurrency.lockutils [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "973ac4ab-5f82-4108-8076-9a370ca9d914" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.015154] env[62692]: DEBUG oslo_concurrency.lockutils [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "refresh_cache-cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1155.015364] env[62692]: DEBUG oslo_concurrency.lockutils [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired lock "refresh_cache-cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.015574] env[62692]: DEBUG nova.network.neutron [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1155.024593] env[62692]: DEBUG oslo_vmware.api [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141958, 'name': ReconfigVM_Task, 'duration_secs': 0.411639} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.025063] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Reconfigured VM instance instance-0000006f to attach disk [datastore1] 45eccffc-f059-4bef-97da-dacb16f20f88/45eccffc-f059-4bef-97da-dacb16f20f88.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1155.030038] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-96ea45c8-8699-4aac-a382-32fe5ebb75b3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.037036] env[62692]: DEBUG oslo_vmware.api [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1155.037036] env[62692]: value = "task-1141959" [ 1155.037036] env[62692]: _type = "Task" [ 1155.037036] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.044922] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a57dbad-7462-40ec-93b7-254edd3d217a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.052052] env[62692]: DEBUG oslo_vmware.api [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141959, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.058202] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8833636-7efc-4d96-8719-ac5156aecfc1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.102598] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd09b2c9-bfac-4568-930e-6043c4eca736 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.114888] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-784aac5b-7d2e-44be-a3cc-1465f15c4e45 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.131804] env[62692]: DEBUG nova.compute.provider_tree [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1155.143726] env[62692]: DEBUG oslo_vmware.api [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141956, 'name': CloneVM_Task} progress is 95%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.285751] env[62692]: DEBUG oslo_vmware.api [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141948, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.369870] env[62692]: INFO nova.compute.manager [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Rescuing [ 1155.370259] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquiring lock "refresh_cache-2b21e8a1-5f5e-4f1f-bae6-512f1213eb48" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1155.370259] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquired lock "refresh_cache-2b21e8a1-5f5e-4f1f-bae6-512f1213eb48" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.370259] env[62692]: DEBUG nova.network.neutron [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1155.390647] env[62692]: INFO nova.compute.manager [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Took 14.12 seconds to build instance. [ 1155.513957] env[62692]: DEBUG nova.network.neutron [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Updating instance_info_cache with network_info: [{"id": "a291ddc4-8996-4755-b347-066f38e6ec7d", "address": "fa:16:3e:7b:10:2a", "network": {"id": "2febb660-329d-403c-8c2c-f11c0f5f547c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1806650357-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37921f67664544478cb2568def068b11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa291ddc4-89", "ovs_interfaceid": "a291ddc4-8996-4755-b347-066f38e6ec7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1155.519621] env[62692]: DEBUG nova.compute.manager [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1155.548803] env[62692]: DEBUG oslo_vmware.api [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141959, 'name': Rename_Task, 'duration_secs': 0.197592} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.549192] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1155.549505] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-24af7cc4-8e91-4515-a5ca-e21b00ed0a86 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.559531] env[62692]: DEBUG oslo_vmware.api [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1155.559531] env[62692]: value = "task-1141960" [ 1155.559531] env[62692]: _type = "Task" [ 1155.559531] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.569165] env[62692]: DEBUG oslo_vmware.api [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141960, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.641512] env[62692]: DEBUG nova.scheduler.client.report [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1155.652285] env[62692]: DEBUG oslo_vmware.api [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141956, 'name': CloneVM_Task, 'duration_secs': 1.704141} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.652688] env[62692]: INFO nova.virt.vmwareapi.vmops [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Created linked-clone VM from snapshot [ 1155.655129] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e45c9df9-1660-43d2-8050-54585f7f03af {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.664592] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Uploading image a661e426-af79-4233-8845-0c9ddc6538a5 {{(pid=62692) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1155.694285] env[62692]: DEBUG oslo_vmware.rw_handles [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1155.694285] env[62692]: value = "vm-249075" [ 1155.694285] env[62692]: _type = "VirtualMachine" [ 1155.694285] env[62692]: }. {{(pid=62692) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1155.694633] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-38e00304-3608-455b-90c2-abf3d3c3bcab {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.705443] env[62692]: DEBUG oslo_vmware.rw_handles [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lease: (returnval){ [ 1155.705443] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5237f291-ae9d-9e26-9c9b-504cb21fcca9" [ 1155.705443] env[62692]: _type = "HttpNfcLease" [ 1155.705443] env[62692]: } obtained for exporting VM: (result){ [ 1155.705443] env[62692]: value = "vm-249075" [ 1155.705443] env[62692]: _type = "VirtualMachine" [ 1155.705443] env[62692]: }. {{(pid=62692) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1155.705731] env[62692]: DEBUG oslo_vmware.api [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the lease: (returnval){ [ 1155.705731] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5237f291-ae9d-9e26-9c9b-504cb21fcca9" [ 1155.705731] env[62692]: _type = "HttpNfcLease" [ 1155.705731] env[62692]: } to be ready. {{(pid=62692) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1155.717962] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1155.717962] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5237f291-ae9d-9e26-9c9b-504cb21fcca9" [ 1155.717962] env[62692]: _type = "HttpNfcLease" [ 1155.717962] env[62692]: } is initializing. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1155.787454] env[62692]: DEBUG oslo_vmware.api [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141948, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.810060] env[62692]: DEBUG nova.network.neutron [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Updating instance_info_cache with network_info: [{"id": "a3885c51-6f7b-41e8-8948-41fac442a288", "address": "fa:16:3e:e4:9b:6f", "network": {"id": "f70727d6-a9aa-44b0-a244-2533b3f96571", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-284105837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1037e5d436bd429391159b87a712f6e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3885c51-6f", "ovs_interfaceid": "a3885c51-6f7b-41e8-8948-41fac442a288", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1155.892273] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc793ead-d11f-4ca0-9f4a-1e155e317c30 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "0b7ea87c-252a-48b3-a1cd-c81e91362812" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.633s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1156.016536] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Releasing lock "refresh_cache-e2383a6a-3581-40fc-a0eb-6981acdbf54a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1156.016774] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Updated the network info_cache for instance {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1156.017045] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1156.017223] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1156.017372] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1156.017521] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1156.017903] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1156.017903] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1156.018086] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62692) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1156.018894] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager.update_available_resource {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1156.044681] env[62692]: DEBUG oslo_concurrency.lockutils [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1156.071807] env[62692]: DEBUG oslo_vmware.api [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141960, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.128242] env[62692]: DEBUG nova.network.neutron [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Updating instance_info_cache with network_info: [{"id": "144e96aa-d02a-48e4-a91e-73909db5c202", "address": "fa:16:3e:23:7e:07", "network": {"id": "50c1330e-1a5f-42d1-9987-79ddcc0ad7aa", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-813901268-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b1c44eb72b334ee3add29026d312e37b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f92f0b92-d6fb-4d00-8ad5-6b3809ed5493", "external-id": "nsx-vlan-transportzone-48", "segmentation_id": 48, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap144e96aa-d0", "ovs_interfaceid": "144e96aa-d02a-48e4-a91e-73909db5c202", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1156.215282] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1156.215282] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5237f291-ae9d-9e26-9c9b-504cb21fcca9" [ 1156.215282] env[62692]: _type = "HttpNfcLease" [ 1156.215282] env[62692]: } is ready. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1156.215822] env[62692]: DEBUG oslo_vmware.rw_handles [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1156.215822] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5237f291-ae9d-9e26-9c9b-504cb21fcca9" [ 1156.215822] env[62692]: _type = "HttpNfcLease" [ 1156.215822] env[62692]: }. {{(pid=62692) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1156.216384] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7c8d39d-c46e-4b2c-a2fa-50d1f067e684 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.225195] env[62692]: DEBUG oslo_vmware.rw_handles [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525e3a00-51a7-6640-0ddd-1f99f5092c6f/disk-0.vmdk from lease info. {{(pid=62692) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1156.225446] env[62692]: DEBUG oslo_vmware.rw_handles [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525e3a00-51a7-6640-0ddd-1f99f5092c6f/disk-0.vmdk for reading. {{(pid=62692) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1156.291753] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquiring lock "8abacc6f-054f-4c4f-abb2-83f20748cf1c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1156.292260] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "8abacc6f-054f-4c4f-abb2-83f20748cf1c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1156.301819] env[62692]: DEBUG oslo_vmware.api [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141948, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.316910] env[62692]: DEBUG oslo_concurrency.lockutils [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Releasing lock "refresh_cache-cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1156.317506] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1156.317677] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-db0df523-39e5-450c-b962-d50eb8b17a97 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.325998] env[62692]: DEBUG oslo_vmware.api [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1156.325998] env[62692]: value = "task-1141962" [ 1156.325998] env[62692]: _type = "Task" [ 1156.325998] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.336777] env[62692]: DEBUG oslo_vmware.api [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141962, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.357556] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-a2f0792d-0ed7-4282-93d5-6e7936d01e64 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.522032] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1156.573464] env[62692]: DEBUG oslo_vmware.api [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1141960, 'name': PowerOnVM_Task, 'duration_secs': 0.543804} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.573768] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1156.573982] env[62692]: INFO nova.compute.manager [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Took 8.45 seconds to spawn the instance on the hypervisor. [ 1156.574218] env[62692]: DEBUG nova.compute.manager [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1156.575154] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd9c82a-62f4-42d6-8f84-02bd1639139d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.629761] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Releasing lock "refresh_cache-2b21e8a1-5f5e-4f1f-bae6-512f1213eb48" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1156.655557] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.394s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1156.663415] env[62692]: DEBUG oslo_concurrency.lockutils [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.619s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1156.665215] env[62692]: INFO nova.compute.claims [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1156.795057] env[62692]: DEBUG nova.compute.manager [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1156.800971] env[62692]: DEBUG nova.compute.manager [req-ce92c4b8-99f4-4eaf-bb8e-3f31077f09ce req-53243827-7d55-4984-b758-1761a04573cc service nova] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Received event network-changed-34de5e92-0920-4a35-bb28-51c7d09bf35d {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1156.801538] env[62692]: DEBUG nova.compute.manager [req-ce92c4b8-99f4-4eaf-bb8e-3f31077f09ce req-53243827-7d55-4984-b758-1761a04573cc service nova] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Refreshing instance network info cache due to event network-changed-34de5e92-0920-4a35-bb28-51c7d09bf35d. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1156.801705] env[62692]: DEBUG oslo_concurrency.lockutils [req-ce92c4b8-99f4-4eaf-bb8e-3f31077f09ce req-53243827-7d55-4984-b758-1761a04573cc service nova] Acquiring lock "refresh_cache-0b7ea87c-252a-48b3-a1cd-c81e91362812" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1156.803094] env[62692]: DEBUG oslo_concurrency.lockutils [req-ce92c4b8-99f4-4eaf-bb8e-3f31077f09ce req-53243827-7d55-4984-b758-1761a04573cc service nova] Acquired lock "refresh_cache-0b7ea87c-252a-48b3-a1cd-c81e91362812" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.803235] env[62692]: DEBUG nova.network.neutron [req-ce92c4b8-99f4-4eaf-bb8e-3f31077f09ce req-53243827-7d55-4984-b758-1761a04573cc service nova] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Refreshing network info cache for port 34de5e92-0920-4a35-bb28-51c7d09bf35d {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1156.813780] env[62692]: DEBUG oslo_vmware.api [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141948, 'name': ReconfigVM_Task, 'duration_secs': 5.80078} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.814391] env[62692]: DEBUG oslo_concurrency.lockutils [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1156.814760] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Reconfigured VM to detach interface {{(pid=62692) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1156.837652] env[62692]: DEBUG oslo_vmware.api [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141962, 'name': PowerOffVM_Task, 'duration_secs': 0.372854} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.838016] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1156.839224] env[62692]: DEBUG nova.virt.hardware [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1156.839336] env[62692]: DEBUG nova.virt.hardware [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1156.839435] env[62692]: DEBUG nova.virt.hardware [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1156.839890] env[62692]: DEBUG nova.virt.hardware [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1156.840191] env[62692]: DEBUG nova.virt.hardware [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1156.840397] env[62692]: DEBUG nova.virt.hardware [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1156.840843] env[62692]: DEBUG nova.virt.hardware [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1156.840843] env[62692]: DEBUG nova.virt.hardware [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1156.840958] env[62692]: DEBUG nova.virt.hardware [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1156.841112] env[62692]: DEBUG nova.virt.hardware [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1156.841295] env[62692]: DEBUG nova.virt.hardware [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1156.847112] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-71451bb3-192c-4450-83d5-bff15c8abe5a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.867809] env[62692]: DEBUG oslo_vmware.api [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1156.867809] env[62692]: value = "task-1141963" [ 1156.867809] env[62692]: _type = "Task" [ 1156.867809] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.881928] env[62692]: DEBUG oslo_vmware.api [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141963, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.098946] env[62692]: INFO nova.compute.manager [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Took 14.64 seconds to build instance. [ 1157.178787] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1157.178787] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6cf616e4-fd03-43a4-ae93-eb8b2fa5b257 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.188689] env[62692]: DEBUG oslo_vmware.api [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1157.188689] env[62692]: value = "task-1141964" [ 1157.188689] env[62692]: _type = "Task" [ 1157.188689] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.204614] env[62692]: DEBUG oslo_vmware.api [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141964, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.241170] env[62692]: INFO nova.scheduler.client.report [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Deleted allocation for migration 34373c4c-3481-421d-b142-1919d2e93950 [ 1157.323673] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1157.380894] env[62692]: DEBUG oslo_vmware.api [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141963, 'name': ReconfigVM_Task, 'duration_secs': 0.245786} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.382013] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-950b24a7-1f75-40c8-a55b-c964b6059368 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.404117] env[62692]: DEBUG nova.virt.hardware [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1157.404472] env[62692]: DEBUG nova.virt.hardware [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1157.404852] env[62692]: DEBUG nova.virt.hardware [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1157.405369] env[62692]: DEBUG nova.virt.hardware [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1157.405621] env[62692]: DEBUG nova.virt.hardware [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1157.405829] env[62692]: DEBUG nova.virt.hardware [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1157.406138] env[62692]: DEBUG nova.virt.hardware [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1157.406397] env[62692]: DEBUG nova.virt.hardware [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1157.406708] env[62692]: DEBUG nova.virt.hardware [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1157.407141] env[62692]: DEBUG nova.virt.hardware [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1157.407256] env[62692]: DEBUG nova.virt.hardware [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1157.408182] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4840018c-01b6-412b-96df-b23acf9a0a51 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.420349] env[62692]: DEBUG oslo_vmware.api [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1157.420349] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]528d043a-5e1a-6e2a-6964-9dc130fe370a" [ 1157.420349] env[62692]: _type = "Task" [ 1157.420349] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.432097] env[62692]: DEBUG oslo_vmware.api [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]528d043a-5e1a-6e2a-6964-9dc130fe370a, 'name': SearchDatastore_Task, 'duration_secs': 0.013458} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.438539] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Reconfiguring VM instance instance-00000062 to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1157.439108] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fee142b6-e21e-4f0f-b4f1-1280834e58f5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.461100] env[62692]: DEBUG oslo_vmware.api [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1157.461100] env[62692]: value = "task-1141965" [ 1157.461100] env[62692]: _type = "Task" [ 1157.461100] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.476239] env[62692]: DEBUG oslo_vmware.api [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141965, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.601653] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b86182b5-f441-40d3-94c6-f336195dd704 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "45eccffc-f059-4bef-97da-dacb16f20f88" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.156s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1157.610464] env[62692]: DEBUG nova.network.neutron [req-ce92c4b8-99f4-4eaf-bb8e-3f31077f09ce req-53243827-7d55-4984-b758-1761a04573cc service nova] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Updated VIF entry in instance network info cache for port 34de5e92-0920-4a35-bb28-51c7d09bf35d. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1157.611123] env[62692]: DEBUG nova.network.neutron [req-ce92c4b8-99f4-4eaf-bb8e-3f31077f09ce req-53243827-7d55-4984-b758-1761a04573cc service nova] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Updating instance_info_cache with network_info: [{"id": "34de5e92-0920-4a35-bb28-51c7d09bf35d", "address": "fa:16:3e:b7:f9:bc", "network": {"id": "e9c538a3-7723-42cb-85b7-5a1597e49cef", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1878627765-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.130", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71b00c566cea4b21811ea1e1ef5ba0f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34de5e92-09", "ovs_interfaceid": "34de5e92-0920-4a35-bb28-51c7d09bf35d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1157.705806] env[62692]: DEBUG oslo_vmware.api [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141964, 'name': PowerOffVM_Task, 'duration_secs': 0.314272} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.709425] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1157.711591] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-243c5094-4400-4a0c-8be0-2060da1a48ee {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.737182] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caa979eb-114f-43ad-9a26-24c0adda49ce {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.747201] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "e55a1a5b-3656-48d7-a610-04610b43999a" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.885s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1157.776579] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1157.777105] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-96ed0590-9953-48bb-8045-68dba3d8f384 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.788244] env[62692]: DEBUG oslo_vmware.api [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1157.788244] env[62692]: value = "task-1141966" [ 1157.788244] env[62692]: _type = "Task" [ 1157.788244] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.809985] env[62692]: DEBUG oslo_vmware.api [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141966, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.917536] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00a435ec-c79c-4970-aaa6-0969b4f4a597 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.926740] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f88a1e7-7580-4fed-9811-70b8c323b170 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.968185] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbdc55d6-11cd-489a-9bec-88c7acbab2a7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.977690] env[62692]: DEBUG oslo_vmware.api [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141965, 'name': ReconfigVM_Task, 'duration_secs': 0.245168} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.980253] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Reconfigured VM instance instance-00000062 to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1157.981274] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c80ea24-4347-426b-a32b-25d79136f021 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.985133] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73b432d4-fe63-4f6f-b365-2f6b209e6d09 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.011804] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43/cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1158.021197] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "1904edb2-fd72-43b6-9219-8043c5abff0c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.021621] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "1904edb2-fd72-43b6-9219-8043c5abff0c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.021895] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "1904edb2-fd72-43b6-9219-8043c5abff0c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.022194] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "1904edb2-fd72-43b6-9219-8043c5abff0c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.022389] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "1904edb2-fd72-43b6-9219-8043c5abff0c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.024304] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f05695c3-bb86-4476-97a7-32a1c173c43f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.038067] env[62692]: DEBUG nova.compute.provider_tree [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1158.041163] env[62692]: INFO nova.compute.manager [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Terminating instance [ 1158.042289] env[62692]: DEBUG nova.compute.manager [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1158.042737] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1158.044135] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f675391e-7ba5-4d27-8a24-d8d7b5288c9e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.050156] env[62692]: DEBUG oslo_vmware.api [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1158.050156] env[62692]: value = "task-1141967" [ 1158.050156] env[62692]: _type = "Task" [ 1158.050156] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.056579] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1158.057341] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2d4b4b37-3aba-4361-9989-773832af242a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.063024] env[62692]: DEBUG oslo_vmware.api [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141967, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.064518] env[62692]: DEBUG oslo_vmware.api [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1158.064518] env[62692]: value = "task-1141968" [ 1158.064518] env[62692]: _type = "Task" [ 1158.064518] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.075283] env[62692]: DEBUG oslo_vmware.api [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141968, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.114530] env[62692]: DEBUG oslo_concurrency.lockutils [req-ce92c4b8-99f4-4eaf-bb8e-3f31077f09ce req-53243827-7d55-4984-b758-1761a04573cc service nova] Releasing lock "refresh_cache-0b7ea87c-252a-48b3-a1cd-c81e91362812" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1158.199617] env[62692]: DEBUG oslo_concurrency.lockutils [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "refresh_cache-1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1158.199962] env[62692]: DEBUG oslo_concurrency.lockutils [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquired lock "refresh_cache-1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1158.200186] env[62692]: DEBUG nova.network.neutron [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1158.301182] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] VM already powered off {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1158.301498] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1158.301685] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1158.301870] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1158.302103] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1158.302486] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d5bce828-841d-4646-8bbd-89e94d1545da {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.314702] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1158.314702] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1158.315418] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-433f97a8-8073-4abc-8670-4799bc2db5fc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.322570] env[62692]: DEBUG oslo_vmware.api [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1158.322570] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5278f4e7-e694-8f43-360f-c898f3ce1490" [ 1158.322570] env[62692]: _type = "Task" [ 1158.322570] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.333652] env[62692]: DEBUG oslo_vmware.api [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5278f4e7-e694-8f43-360f-c898f3ce1490, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.542696] env[62692]: DEBUG nova.scheduler.client.report [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1158.564421] env[62692]: DEBUG oslo_vmware.api [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141967, 'name': ReconfigVM_Task, 'duration_secs': 0.442897} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.568024] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Reconfigured VM instance instance-00000062 to attach disk [datastore2] cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43/cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1158.569340] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd537358-6bda-41ed-a52b-0ac803c94a41 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.594175] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab4aca48-05f4-4026-9673-1000934e017b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.597337] env[62692]: DEBUG oslo_vmware.api [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141968, 'name': PowerOffVM_Task, 'duration_secs': 0.189431} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.597890] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1158.598240] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1158.599279] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5eb9613e-4cec-4263-9cb6-e5805e393ee3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.616446] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bd290ea-4573-4f15-9bed-f7f4309a7303 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.637612] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b423118e-971b-4518-8d7b-64cd1f005376 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.648356] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1158.649228] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-678978c3-6c77-424f-aa55-f0ce62129fd2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.658868] env[62692]: DEBUG oslo_vmware.api [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1158.658868] env[62692]: value = "task-1141970" [ 1158.658868] env[62692]: _type = "Task" [ 1158.658868] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.669065] env[62692]: DEBUG oslo_vmware.api [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141970, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.670728] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1158.671131] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1158.671437] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Deleting the datastore file [datastore1] 1904edb2-fd72-43b6-9219-8043c5abff0c {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1158.671826] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a1ee9712-033e-436f-91e5-ebbff12fec15 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.678588] env[62692]: DEBUG oslo_vmware.api [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1158.678588] env[62692]: value = "task-1141971" [ 1158.678588] env[62692]: _type = "Task" [ 1158.678588] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.688295] env[62692]: DEBUG oslo_vmware.api [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141971, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.835557] env[62692]: DEBUG oslo_vmware.api [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5278f4e7-e694-8f43-360f-c898f3ce1490, 'name': SearchDatastore_Task, 'duration_secs': 0.01288} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.837027] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38b89238-15bf-4ab0-b5bf-68fa38d476df {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.845735] env[62692]: DEBUG oslo_vmware.api [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1158.845735] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5284cf68-1551-0c60-3d34-6eb83c221952" [ 1158.845735] env[62692]: _type = "Task" [ 1158.845735] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.853955] env[62692]: DEBUG nova.compute.manager [req-0ecd2582-1bdd-4532-8b93-837ff3d64c11 req-180c348e-25b1-4242-964d-b453fcd26516 service nova] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Received event network-changed-01bf6621-08ff-4983-8bcb-1d178bda2c60 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1158.854402] env[62692]: DEBUG nova.compute.manager [req-0ecd2582-1bdd-4532-8b93-837ff3d64c11 req-180c348e-25b1-4242-964d-b453fcd26516 service nova] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Refreshing instance network info cache due to event network-changed-01bf6621-08ff-4983-8bcb-1d178bda2c60. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1158.854838] env[62692]: DEBUG oslo_concurrency.lockutils [req-0ecd2582-1bdd-4532-8b93-837ff3d64c11 req-180c348e-25b1-4242-964d-b453fcd26516 service nova] Acquiring lock "refresh_cache-45eccffc-f059-4bef-97da-dacb16f20f88" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1158.855189] env[62692]: DEBUG oslo_concurrency.lockutils [req-0ecd2582-1bdd-4532-8b93-837ff3d64c11 req-180c348e-25b1-4242-964d-b453fcd26516 service nova] Acquired lock "refresh_cache-45eccffc-f059-4bef-97da-dacb16f20f88" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1158.855533] env[62692]: DEBUG nova.network.neutron [req-0ecd2582-1bdd-4532-8b93-837ff3d64c11 req-180c348e-25b1-4242-964d-b453fcd26516 service nova] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Refreshing network info cache for port 01bf6621-08ff-4983-8bcb-1d178bda2c60 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1158.866170] env[62692]: DEBUG oslo_vmware.api [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5284cf68-1551-0c60-3d34-6eb83c221952, 'name': SearchDatastore_Task, 'duration_secs': 0.014882} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.869581] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1158.869996] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48/81eb64fb-5a0c-47b7-9948-7dcb01b911bd-rescue.vmdk. {{(pid=62692) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1158.871393] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a77bad15-3052-4ba9-87db-d227967ee398 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.885284] env[62692]: DEBUG oslo_vmware.api [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1158.885284] env[62692]: value = "task-1141972" [ 1158.885284] env[62692]: _type = "Task" [ 1158.885284] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.886172] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "e55a1a5b-3656-48d7-a610-04610b43999a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.886655] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "e55a1a5b-3656-48d7-a610-04610b43999a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.886817] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "e55a1a5b-3656-48d7-a610-04610b43999a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.886875] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "e55a1a5b-3656-48d7-a610-04610b43999a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.887078] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "e55a1a5b-3656-48d7-a610-04610b43999a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.894512] env[62692]: INFO nova.compute.manager [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Terminating instance [ 1158.900021] env[62692]: DEBUG nova.compute.manager [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1158.900021] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1158.900021] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b421ffa4-8d3e-4bf7-853c-aba79389f066 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.907480] env[62692]: DEBUG oslo_vmware.api [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141972, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.914075] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1158.914401] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fc42c852-8a58-4002-8d64-29e64484a145 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.925481] env[62692]: DEBUG oslo_vmware.api [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1158.925481] env[62692]: value = "task-1141973" [ 1158.925481] env[62692]: _type = "Task" [ 1158.925481] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.939040] env[62692]: DEBUG oslo_vmware.api [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141973, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.050944] env[62692]: DEBUG oslo_concurrency.lockutils [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.387s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1159.051240] env[62692]: DEBUG nova.compute.manager [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1159.055885] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.534s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1159.056104] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1159.056442] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62692) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1159.056843] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.733s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1159.059030] env[62692]: INFO nova.compute.claims [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1159.063799] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2af06f3c-33a4-4d7d-9aa7-e20acc7010e1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.069633] env[62692]: INFO nova.network.neutron [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Port 613360e6-9cd1-4423-9c92-293618e0c0d1 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1159.070150] env[62692]: DEBUG nova.network.neutron [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Updating instance_info_cache with network_info: [{"id": "a668d6b6-4111-4a94-8bc8-09535b29f895", "address": "fa:16:3e:02:7f:2a", "network": {"id": "b34af32c-59b7-4de4-85ef-08bd67c1e2a2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1972705585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ba97a775e99a463c845518b0df8cd3e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa668d6b6-41", "ovs_interfaceid": "a668d6b6-4111-4a94-8bc8-09535b29f895", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1159.081734] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57f8d5ef-b7d8-4d52-b1ad-68a6c6cb4050 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.103502] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8afb02cf-1ec7-44e4-a11a-d6c83644499c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.119051] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66f95044-8fa1-431d-92eb-43f1eb1313d4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.175106] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180207MB free_disk=166GB free_vcpus=48 pci_devices=None {{(pid=62692) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1159.175106] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1159.185777] env[62692]: DEBUG oslo_vmware.api [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141970, 'name': PowerOnVM_Task, 'duration_secs': 0.453267} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.186463] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1159.194347] env[62692]: DEBUG oslo_vmware.api [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141971, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.24867} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.194347] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1159.194347] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1159.194347] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1159.194347] env[62692]: INFO nova.compute.manager [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1159.194642] env[62692]: DEBUG oslo.service.loopingcall [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1159.194786] env[62692]: DEBUG nova.compute.manager [-] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1159.194868] env[62692]: DEBUG nova.network.neutron [-] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1159.398584] env[62692]: DEBUG oslo_vmware.api [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141972, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.438351] env[62692]: DEBUG oslo_vmware.api [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141973, 'name': PowerOffVM_Task, 'duration_secs': 0.252737} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.438643] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1159.438817] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1159.439189] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aa0deef8-6d0c-4e2e-a6ec-5b2403513165 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.529988] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1159.530200] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1159.530400] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Deleting the datastore file [datastore1] e55a1a5b-3656-48d7-a610-04610b43999a {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1159.530698] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dfc8253a-c4be-4e58-a7ed-fa8e42c56d12 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.542969] env[62692]: DEBUG oslo_vmware.api [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for the task: (returnval){ [ 1159.542969] env[62692]: value = "task-1141975" [ 1159.542969] env[62692]: _type = "Task" [ 1159.542969] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.553959] env[62692]: DEBUG oslo_vmware.api [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141975, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.564636] env[62692]: DEBUG nova.compute.utils [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1159.566060] env[62692]: DEBUG nova.compute.manager [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1159.566232] env[62692]: DEBUG nova.network.neutron [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1159.575074] env[62692]: DEBUG oslo_concurrency.lockutils [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Releasing lock "refresh_cache-1904edb2-fd72-43b6-9219-8043c5abff0c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1159.657768] env[62692]: DEBUG nova.policy [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '016134ac36fd4bd893226ae95a955e28', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37921f67664544478cb2568def068b11', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1159.691244] env[62692]: DEBUG nova.network.neutron [req-0ecd2582-1bdd-4532-8b93-837ff3d64c11 req-180c348e-25b1-4242-964d-b453fcd26516 service nova] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Updated VIF entry in instance network info cache for port 01bf6621-08ff-4983-8bcb-1d178bda2c60. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1159.691646] env[62692]: DEBUG nova.network.neutron [req-0ecd2582-1bdd-4532-8b93-837ff3d64c11 req-180c348e-25b1-4242-964d-b453fcd26516 service nova] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Updating instance_info_cache with network_info: [{"id": "01bf6621-08ff-4983-8bcb-1d178bda2c60", "address": "fa:16:3e:15:0e:b2", "network": {"id": "26fe2f12-b216-4fa5-a022-24c5ab315416", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-838584488-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb2ee16ec90c48e6babfb1d31ea3de94", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01bf6621-08", "ovs_interfaceid": "01bf6621-08ff-4983-8bcb-1d178bda2c60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1159.904021] env[62692]: DEBUG oslo_vmware.api [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141972, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.760819} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.904021] env[62692]: INFO nova.virt.vmwareapi.ds_util [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48/81eb64fb-5a0c-47b7-9948-7dcb01b911bd-rescue.vmdk. [ 1159.904021] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee5020c7-bd7a-4688-9a48-4df381e21ad3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.930140] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48/81eb64fb-5a0c-47b7-9948-7dcb01b911bd-rescue.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1159.932471] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ade5b34d-9ebd-4ce3-b53e-d18632ab202f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.955392] env[62692]: DEBUG oslo_vmware.api [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1159.955392] env[62692]: value = "task-1141976" [ 1159.955392] env[62692]: _type = "Task" [ 1159.955392] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.967164] env[62692]: DEBUG oslo_vmware.api [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141976, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.012572] env[62692]: DEBUG nova.network.neutron [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Successfully created port: e2bdb2a3-7e2c-4329-9008-9bea98b0b80a {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1160.054887] env[62692]: DEBUG oslo_vmware.api [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Task: {'id': task-1141975, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.490644} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.058018] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1160.058018] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1160.058018] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1160.058018] env[62692]: INFO nova.compute.manager [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1160.058018] env[62692]: DEBUG oslo.service.loopingcall [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1160.058018] env[62692]: DEBUG nova.compute.manager [-] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1160.058018] env[62692]: DEBUG nova.network.neutron [-] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1160.069891] env[62692]: DEBUG nova.compute.manager [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1160.077116] env[62692]: DEBUG oslo_concurrency.lockutils [None req-00dcc68f-00ee-4b76-b913-f1fb864133d7 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "interface-1904edb2-fd72-43b6-9219-8043c5abff0c-613360e6-9cd1-4423-9c92-293618e0c0d1" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.889s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1160.194426] env[62692]: DEBUG oslo_concurrency.lockutils [req-0ecd2582-1bdd-4532-8b93-837ff3d64c11 req-180c348e-25b1-4242-964d-b453fcd26516 service nova] Releasing lock "refresh_cache-45eccffc-f059-4bef-97da-dacb16f20f88" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1160.199022] env[62692]: INFO nova.compute.manager [None req-341d0370-b610-4253-a3bf-9797147b691a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Updating instance to original state: 'active' [ 1160.340737] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bed6f3d-d0d6-4e4b-9570-765022200f1e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.351147] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dea8bfb7-bbc4-49a8-83b6-0ff7a9f3855b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.387697] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a6d6a2-3ade-4a21-9568-15fbcdf79092 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.397637] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73aaf522-e27f-43b0-b1c3-b43b88caec37 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.418034] env[62692]: DEBUG nova.compute.provider_tree [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1160.448438] env[62692]: DEBUG nova.network.neutron [-] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1160.467223] env[62692]: DEBUG oslo_vmware.api [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141976, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.581395] env[62692]: INFO nova.virt.block_device [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Booting with volume d3635284-5881-414b-8d43-075170b6ad32 at /dev/sda [ 1160.619483] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fc9a9ec8-0318-4a39-8458-a1c305fec550 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.632796] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-736c87d9-4dc4-46b6-85b2-b292242f8943 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.668413] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ced652df-e129-4f14-a864-f2ee986da24e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.678099] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-575c19a8-9905-4d4a-9fd5-abc59e900da4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.714992] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6baf7191-9bfe-4071-a0ea-ccef3dcdde8b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.729029] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-269736f8-2664-479c-8422-af8317a6b653 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.742960] env[62692]: DEBUG nova.virt.block_device [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Updating existing volume attachment record: e9bd0fc1-d683-4faa-a377-fb15a3882af9 {{(pid=62692) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1160.890853] env[62692]: DEBUG nova.compute.manager [req-50d15f08-e5cd-4e51-a575-d28e4557c627 req-7e42f4e2-e255-427d-8c25-372988edb79e service nova] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Received event network-vif-deleted-a668d6b6-4111-4a94-8bc8-09535b29f895 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1160.891116] env[62692]: DEBUG nova.compute.manager [req-50d15f08-e5cd-4e51-a575-d28e4557c627 req-7e42f4e2-e255-427d-8c25-372988edb79e service nova] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Received event network-vif-deleted-5e50f271-31a9-4755-a612-6b971179167a {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1160.891309] env[62692]: INFO nova.compute.manager [req-50d15f08-e5cd-4e51-a575-d28e4557c627 req-7e42f4e2-e255-427d-8c25-372988edb79e service nova] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Neutron deleted interface 5e50f271-31a9-4755-a612-6b971179167a; detaching it from the instance and deleting it from the info cache [ 1160.891491] env[62692]: DEBUG nova.network.neutron [req-50d15f08-e5cd-4e51-a575-d28e4557c627 req-7e42f4e2-e255-427d-8c25-372988edb79e service nova] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1160.892895] env[62692]: DEBUG nova.network.neutron [-] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1160.921082] env[62692]: DEBUG nova.scheduler.client.report [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1160.951511] env[62692]: INFO nova.compute.manager [-] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Took 1.76 seconds to deallocate network for instance. [ 1160.973907] env[62692]: DEBUG oslo_vmware.api [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141976, 'name': ReconfigVM_Task, 'duration_secs': 0.881754} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.974355] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48/81eb64fb-5a0c-47b7-9948-7dcb01b911bd-rescue.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1160.976117] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a957dbb-c9c4-41a1-9b61-d6b9845e0a13 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.018697] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ee680a9c-e342-4528-8def-c72f0914e520 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.039416] env[62692]: DEBUG oslo_vmware.api [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1161.039416] env[62692]: value = "task-1141977" [ 1161.039416] env[62692]: _type = "Task" [ 1161.039416] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.048337] env[62692]: DEBUG oslo_vmware.api [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141977, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.396031] env[62692]: INFO nova.compute.manager [-] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Took 1.34 seconds to deallocate network for instance. [ 1161.396031] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-735bd22e-d091-4162-a661-72c9ed713e68 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.409048] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b16e0c09-01d6-48f9-953c-ddd73ce1a6f6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.425607] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.369s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1161.426238] env[62692]: DEBUG nova.compute.manager [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1161.429618] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 2.255s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1161.446113] env[62692]: DEBUG nova.compute.manager [req-50d15f08-e5cd-4e51-a575-d28e4557c627 req-7e42f4e2-e255-427d-8c25-372988edb79e service nova] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Detach interface failed, port_id=5e50f271-31a9-4755-a612-6b971179167a, reason: Instance e55a1a5b-3656-48d7-a610-04610b43999a could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1161.469951] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1161.551035] env[62692]: DEBUG oslo_vmware.api [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141977, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.714219] env[62692]: DEBUG nova.compute.manager [req-4822137e-b3a2-4b0e-b4a9-0f9913167ea0 req-9dea2005-d99c-43a7-9c3b-7452f3ed66f1 service nova] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Received event network-vif-plugged-e2bdb2a3-7e2c-4329-9008-9bea98b0b80a {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1161.714502] env[62692]: DEBUG oslo_concurrency.lockutils [req-4822137e-b3a2-4b0e-b4a9-0f9913167ea0 req-9dea2005-d99c-43a7-9c3b-7452f3ed66f1 service nova] Acquiring lock "973ac4ab-5f82-4108-8076-9a370ca9d914-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1161.714651] env[62692]: DEBUG oslo_concurrency.lockutils [req-4822137e-b3a2-4b0e-b4a9-0f9913167ea0 req-9dea2005-d99c-43a7-9c3b-7452f3ed66f1 service nova] Lock "973ac4ab-5f82-4108-8076-9a370ca9d914-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1161.714961] env[62692]: DEBUG oslo_concurrency.lockutils [req-4822137e-b3a2-4b0e-b4a9-0f9913167ea0 req-9dea2005-d99c-43a7-9c3b-7452f3ed66f1 service nova] Lock "973ac4ab-5f82-4108-8076-9a370ca9d914-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1161.715121] env[62692]: DEBUG nova.compute.manager [req-4822137e-b3a2-4b0e-b4a9-0f9913167ea0 req-9dea2005-d99c-43a7-9c3b-7452f3ed66f1 service nova] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] No waiting events found dispatching network-vif-plugged-e2bdb2a3-7e2c-4329-9008-9bea98b0b80a {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1161.715327] env[62692]: WARNING nova.compute.manager [req-4822137e-b3a2-4b0e-b4a9-0f9913167ea0 req-9dea2005-d99c-43a7-9c3b-7452f3ed66f1 service nova] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Received unexpected event network-vif-plugged-e2bdb2a3-7e2c-4329-9008-9bea98b0b80a for instance with vm_state building and task_state block_device_mapping. [ 1161.777305] env[62692]: DEBUG nova.network.neutron [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Successfully updated port: e2bdb2a3-7e2c-4329-9008-9bea98b0b80a {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1161.922552] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1161.933348] env[62692]: DEBUG nova.compute.utils [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1161.941648] env[62692]: DEBUG nova.compute.manager [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1161.941876] env[62692]: DEBUG nova.network.neutron [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1161.982212] env[62692]: DEBUG nova.policy [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '847654db6b294a828455ae22db82c50e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f94f7500bf524e3890734a58937b80c4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1162.051459] env[62692]: DEBUG oslo_vmware.api [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141977, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.245390] env[62692]: DEBUG nova.network.neutron [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Successfully created port: d463b030-eded-4099-ae6f-2aea2d1af078 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1162.283860] env[62692]: DEBUG oslo_concurrency.lockutils [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "refresh_cache-973ac4ab-5f82-4108-8076-9a370ca9d914" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1162.283860] env[62692]: DEBUG oslo_concurrency.lockutils [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquired lock "refresh_cache-973ac4ab-5f82-4108-8076-9a370ca9d914" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1162.283860] env[62692]: DEBUG nova.network.neutron [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1162.283860] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1162.283860] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1162.283860] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1162.283860] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1162.283860] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1162.286823] env[62692]: INFO nova.compute.manager [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Terminating instance [ 1162.288527] env[62692]: DEBUG nova.compute.manager [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1162.288725] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1162.289567] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eba73cab-dba6-4b34-98b2-0a97f68d3e2a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.298597] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1162.298846] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-74fe62da-8fc4-495e-b40a-b55515ab1153 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.306952] env[62692]: DEBUG oslo_vmware.api [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1162.306952] env[62692]: value = "task-1141978" [ 1162.306952] env[62692]: _type = "Task" [ 1162.306952] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.315356] env[62692]: DEBUG oslo_vmware.api [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141978, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.450372] env[62692]: DEBUG nova.compute.manager [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1162.484231] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance e2383a6a-3581-40fc-a0eb-6981acdbf54a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1162.484231] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 90932f05-fad3-4c6d-87ae-ab059351b0be actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1162.484231] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 94a07505-661f-4595-9024-8a9da9f2c2fc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1162.484231] env[62692]: WARNING nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 1904edb2-fd72-43b6-9219-8043c5abff0c is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1162.484231] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance dd8a0d53-5dc8-4b66-add7-f244e600468d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1162.484231] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance c2f2a485-363e-41af-9b3f-37092947ca08 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1162.484231] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 1c7ce408-6e9c-43bd-8d81-6acb12353ed8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1162.485317] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1162.485712] env[62692]: WARNING nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance e55a1a5b-3656-48d7-a610-04610b43999a is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1162.485989] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1162.486257] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 0b7ea87c-252a-48b3-a1cd-c81e91362812 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1162.486486] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 45eccffc-f059-4bef-97da-dacb16f20f88 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1162.486710] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 973ac4ab-5f82-4108-8076-9a370ca9d914 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1162.486934] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 8abacc6f-054f-4c4f-abb2-83f20748cf1c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1162.487254] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Total usable vcpus: 48, total allocated vcpus: 12 {{(pid=62692) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1162.487505] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2816MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=12 pci_stats=[] {{(pid=62692) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1162.557015] env[62692]: DEBUG oslo_vmware.api [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141977, 'name': ReconfigVM_Task, 'duration_secs': 1.350379} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.557081] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1162.557388] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-da00057b-2dd7-4010-abbe-f64de2950953 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.567201] env[62692]: DEBUG oslo_vmware.api [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1162.567201] env[62692]: value = "task-1141979" [ 1162.567201] env[62692]: _type = "Task" [ 1162.567201] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.577485] env[62692]: DEBUG oslo_vmware.api [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141979, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.725012] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-417809f0-11f4-45a6-95e4-3c5bed489dbc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.735935] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93feddb8-2f79-4d90-8195-fb4c509d11f1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.772921] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52850ce0-ef3d-4028-9997-4b8365ec425d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.782654] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18c99e66-8bbd-471c-a0fb-1c0898d45784 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.801312] env[62692]: DEBUG nova.compute.provider_tree [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1162.818346] env[62692]: DEBUG oslo_vmware.api [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141978, 'name': PowerOffVM_Task, 'duration_secs': 0.301949} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.818627] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1162.818823] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1162.819117] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1576b040-308d-4656-bc0b-8f31c01c86c5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.829768] env[62692]: DEBUG nova.network.neutron [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1162.839973] env[62692]: DEBUG nova.compute.manager [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1162.840278] env[62692]: DEBUG nova.virt.hardware [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1162.840503] env[62692]: DEBUG nova.virt.hardware [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1162.840660] env[62692]: DEBUG nova.virt.hardware [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1162.840848] env[62692]: DEBUG nova.virt.hardware [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1162.841008] env[62692]: DEBUG nova.virt.hardware [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1162.841170] env[62692]: DEBUG nova.virt.hardware [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1162.841387] env[62692]: DEBUG nova.virt.hardware [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1162.841553] env[62692]: DEBUG nova.virt.hardware [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1162.841724] env[62692]: DEBUG nova.virt.hardware [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1162.842015] env[62692]: DEBUG nova.virt.hardware [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1162.842347] env[62692]: DEBUG nova.virt.hardware [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1162.843454] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39adbd87-c666-4d69-a74c-46f31dcfe15b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.857089] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f1682a7-39e5-4226-8d2e-7cb64205fb2a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.895888] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1162.896208] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1162.897045] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Deleting the datastore file [datastore2] cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1162.897045] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1723bac6-502a-458c-aed5-9b0047abdfe4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.904733] env[62692]: DEBUG oslo_vmware.api [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1162.904733] env[62692]: value = "task-1141981" [ 1162.904733] env[62692]: _type = "Task" [ 1162.904733] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.917705] env[62692]: DEBUG oslo_vmware.api [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141981, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.000544] env[62692]: DEBUG nova.network.neutron [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Updating instance_info_cache with network_info: [{"id": "e2bdb2a3-7e2c-4329-9008-9bea98b0b80a", "address": "fa:16:3e:3c:8b:ce", "network": {"id": "2febb660-329d-403c-8c2c-f11c0f5f547c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1806650357-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37921f67664544478cb2568def068b11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape2bdb2a3-7e", "ovs_interfaceid": "e2bdb2a3-7e2c-4329-9008-9bea98b0b80a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1163.080730] env[62692]: DEBUG oslo_vmware.api [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141979, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.304390] env[62692]: DEBUG nova.scheduler.client.report [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1163.421293] env[62692]: DEBUG oslo_vmware.api [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1141981, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.283438} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.421596] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1163.421748] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1163.421947] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1163.422160] env[62692]: INFO nova.compute.manager [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1163.422465] env[62692]: DEBUG oslo.service.loopingcall [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1163.422673] env[62692]: DEBUG nova.compute.manager [-] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1163.422769] env[62692]: DEBUG nova.network.neutron [-] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1163.459895] env[62692]: DEBUG nova.compute.manager [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1163.488595] env[62692]: DEBUG nova.virt.hardware [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1163.488890] env[62692]: DEBUG nova.virt.hardware [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1163.489083] env[62692]: DEBUG nova.virt.hardware [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1163.489303] env[62692]: DEBUG nova.virt.hardware [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1163.489462] env[62692]: DEBUG nova.virt.hardware [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1163.489615] env[62692]: DEBUG nova.virt.hardware [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1163.489850] env[62692]: DEBUG nova.virt.hardware [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1163.490036] env[62692]: DEBUG nova.virt.hardware [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1163.490223] env[62692]: DEBUG nova.virt.hardware [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1163.490406] env[62692]: DEBUG nova.virt.hardware [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1163.490588] env[62692]: DEBUG nova.virt.hardware [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1163.491514] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2815090b-45e7-4c35-896a-8616c722efbf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.500904] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a379d43-c3a1-46ee-9d6a-62661555dd4e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.504748] env[62692]: DEBUG oslo_concurrency.lockutils [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Releasing lock "refresh_cache-973ac4ab-5f82-4108-8076-9a370ca9d914" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1163.505355] env[62692]: DEBUG nova.compute.manager [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Instance network_info: |[{"id": "e2bdb2a3-7e2c-4329-9008-9bea98b0b80a", "address": "fa:16:3e:3c:8b:ce", "network": {"id": "2febb660-329d-403c-8c2c-f11c0f5f547c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1806650357-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37921f67664544478cb2568def068b11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape2bdb2a3-7e", "ovs_interfaceid": "e2bdb2a3-7e2c-4329-9008-9bea98b0b80a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1163.505782] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3c:8b:ce', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e2bdb2a3-7e2c-4329-9008-9bea98b0b80a', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1163.513512] env[62692]: DEBUG oslo.service.loopingcall [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1163.514201] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1163.514447] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-176c2bff-d866-4229-9caf-c31dab223ec0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.543954] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1163.543954] env[62692]: value = "task-1141982" [ 1163.543954] env[62692]: _type = "Task" [ 1163.543954] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.552692] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141982, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.580606] env[62692]: DEBUG oslo_vmware.api [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141979, 'name': PowerOnVM_Task, 'duration_secs': 0.573343} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.581048] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1163.584476] env[62692]: DEBUG nova.compute.manager [None req-f9b043d9-50a2-4114-ae4e-6622dcfed44a tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1163.585458] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e690f239-b79b-4231-982c-1c718d2ec18f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.747125] env[62692]: DEBUG nova.compute.manager [req-cf126989-e0b8-407d-a75d-6ce3b0c4128a req-ef0707dc-85e3-4657-af93-f81ce84e302b service nova] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Received event network-changed-e2bdb2a3-7e2c-4329-9008-9bea98b0b80a {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1163.747382] env[62692]: DEBUG nova.compute.manager [req-cf126989-e0b8-407d-a75d-6ce3b0c4128a req-ef0707dc-85e3-4657-af93-f81ce84e302b service nova] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Refreshing instance network info cache due to event network-changed-e2bdb2a3-7e2c-4329-9008-9bea98b0b80a. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1163.747579] env[62692]: DEBUG oslo_concurrency.lockutils [req-cf126989-e0b8-407d-a75d-6ce3b0c4128a req-ef0707dc-85e3-4657-af93-f81ce84e302b service nova] Acquiring lock "refresh_cache-973ac4ab-5f82-4108-8076-9a370ca9d914" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1163.747755] env[62692]: DEBUG oslo_concurrency.lockutils [req-cf126989-e0b8-407d-a75d-6ce3b0c4128a req-ef0707dc-85e3-4657-af93-f81ce84e302b service nova] Acquired lock "refresh_cache-973ac4ab-5f82-4108-8076-9a370ca9d914" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1163.747882] env[62692]: DEBUG nova.network.neutron [req-cf126989-e0b8-407d-a75d-6ce3b0c4128a req-ef0707dc-85e3-4657-af93-f81ce84e302b service nova] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Refreshing network info cache for port e2bdb2a3-7e2c-4329-9008-9bea98b0b80a {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1163.809737] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62692) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1163.810083] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.380s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.810511] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.344s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.810831] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.814993] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.891s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.815297] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.002s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.817452] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1163.843695] env[62692]: INFO nova.scheduler.client.report [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Deleted allocations for instance e55a1a5b-3656-48d7-a610-04610b43999a [ 1163.849943] env[62692]: INFO nova.scheduler.client.report [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Deleted allocations for instance 1904edb2-fd72-43b6-9219-8043c5abff0c [ 1164.042175] env[62692]: DEBUG nova.network.neutron [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Successfully updated port: d463b030-eded-4099-ae6f-2aea2d1af078 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1164.054631] env[62692]: DEBUG nova.compute.manager [req-f00b33d7-25f3-49b7-9fd4-a64fb540748d req-246bd35b-926c-4327-bcef-19b4c892ec23 service nova] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Received event network-vif-deleted-a3885c51-6f7b-41e8-8948-41fac442a288 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1164.055464] env[62692]: INFO nova.compute.manager [req-f00b33d7-25f3-49b7-9fd4-a64fb540748d req-246bd35b-926c-4327-bcef-19b4c892ec23 service nova] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Neutron deleted interface a3885c51-6f7b-41e8-8948-41fac442a288; detaching it from the instance and deleting it from the info cache [ 1164.055789] env[62692]: DEBUG nova.network.neutron [req-f00b33d7-25f3-49b7-9fd4-a64fb540748d req-246bd35b-926c-4327-bcef-19b4c892ec23 service nova] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1164.061822] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141982, 'name': CreateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.365870] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ff2cffe3-337b-494d-8459-2edac921788a tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "1904edb2-fd72-43b6-9219-8043c5abff0c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.344s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.376245] env[62692]: DEBUG oslo_concurrency.lockutils [None req-d6fb3e13-bc44-4ebf-90d9-a7eb8523fbe2 tempest-DeleteServersTestJSON-1426553662 tempest-DeleteServersTestJSON-1426553662-project-member] Lock "e55a1a5b-3656-48d7-a610-04610b43999a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.489s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.488844] env[62692]: DEBUG nova.network.neutron [-] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1164.508328] env[62692]: INFO nova.compute.manager [None req-4c5dba13-620e-4a18-a830-3de73c7e81c6 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Unrescuing [ 1164.508609] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4c5dba13-620e-4a18-a830-3de73c7e81c6 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquiring lock "refresh_cache-2b21e8a1-5f5e-4f1f-bae6-512f1213eb48" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1164.508769] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4c5dba13-620e-4a18-a830-3de73c7e81c6 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquired lock "refresh_cache-2b21e8a1-5f5e-4f1f-bae6-512f1213eb48" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1164.508942] env[62692]: DEBUG nova.network.neutron [None req-4c5dba13-620e-4a18-a830-3de73c7e81c6 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1164.547762] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquiring lock "refresh_cache-8abacc6f-054f-4c4f-abb2-83f20748cf1c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1164.547762] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquired lock "refresh_cache-8abacc6f-054f-4c4f-abb2-83f20748cf1c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1164.547762] env[62692]: DEBUG nova.network.neutron [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1164.560335] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141982, 'name': CreateVM_Task, 'duration_secs': 0.597219} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.561187] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1164.561890] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'mount_device': '/dev/sda', 'attachment_id': 'e9bd0fc1-d683-4faa-a377-fb15a3882af9', 'guest_format': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249070', 'volume_id': 'd3635284-5881-414b-8d43-075170b6ad32', 'name': 'volume-d3635284-5881-414b-8d43-075170b6ad32', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '973ac4ab-5f82-4108-8076-9a370ca9d914', 'attached_at': '', 'detached_at': '', 'volume_id': 'd3635284-5881-414b-8d43-075170b6ad32', 'serial': 'd3635284-5881-414b-8d43-075170b6ad32'}, 'device_type': None, 'delete_on_termination': True, 'disk_bus': None, 'boot_index': 0, 'volume_type': None}], 'swap': None} {{(pid=62692) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1164.562401] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Root volume attach. Driver type: vmdk {{(pid=62692) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1164.564503] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcb75a2b-554b-4969-af82-e23307458b8f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.568103] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-41634484-4b5e-44dc-9ca3-e3cf6b2e156e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.576915] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c333681b-9d22-4382-a6bf-c12c94acd10d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.582587] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc03a49-5309-4e4f-b2cf-44aefee25770 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.593627] env[62692]: DEBUG nova.network.neutron [req-cf126989-e0b8-407d-a75d-6ce3b0c4128a req-ef0707dc-85e3-4657-af93-f81ce84e302b service nova] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Updated VIF entry in instance network info cache for port e2bdb2a3-7e2c-4329-9008-9bea98b0b80a. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1164.593883] env[62692]: DEBUG nova.network.neutron [req-cf126989-e0b8-407d-a75d-6ce3b0c4128a req-ef0707dc-85e3-4657-af93-f81ce84e302b service nova] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Updating instance_info_cache with network_info: [{"id": "e2bdb2a3-7e2c-4329-9008-9bea98b0b80a", "address": "fa:16:3e:3c:8b:ce", "network": {"id": "2febb660-329d-403c-8c2c-f11c0f5f547c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1806650357-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37921f67664544478cb2568def068b11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape2bdb2a3-7e", "ovs_interfaceid": "e2bdb2a3-7e2c-4329-9008-9bea98b0b80a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1164.602080] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42ff0195-0c15-4db7-9b4a-1c2e37c4d6ac {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.619733] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-870f54e5-a3cf-493d-abae-822ef6532658 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.622694] env[62692]: DEBUG nova.compute.manager [req-f00b33d7-25f3-49b7-9fd4-a64fb540748d req-246bd35b-926c-4327-bcef-19b4c892ec23 service nova] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Detach interface failed, port_id=a3885c51-6f7b-41e8-8948-41fac442a288, reason: Instance cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1164.629679] env[62692]: DEBUG oslo_vmware.api [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1164.629679] env[62692]: value = "task-1141984" [ 1164.629679] env[62692]: _type = "Task" [ 1164.629679] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.638721] env[62692]: DEBUG oslo_vmware.api [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141984, 'name': RelocateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.878654] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "94a07505-661f-4595-9024-8a9da9f2c2fc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1164.879069] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "94a07505-661f-4595-9024-8a9da9f2c2fc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1164.879135] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "94a07505-661f-4595-9024-8a9da9f2c2fc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1164.879333] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "94a07505-661f-4595-9024-8a9da9f2c2fc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1164.879512] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "94a07505-661f-4595-9024-8a9da9f2c2fc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.882166] env[62692]: INFO nova.compute.manager [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Terminating instance [ 1164.884295] env[62692]: DEBUG nova.compute.manager [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1164.884513] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1164.885481] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b2afec2-08c1-473a-89fc-e27de49a5741 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.893672] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1164.893947] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7d28c8f2-9f59-467b-b598-d5df4aa9e73c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.901320] env[62692]: DEBUG oslo_vmware.api [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1164.901320] env[62692]: value = "task-1141985" [ 1164.901320] env[62692]: _type = "Task" [ 1164.901320] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.911367] env[62692]: DEBUG oslo_vmware.api [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141985, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.994149] env[62692]: INFO nova.compute.manager [-] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Took 1.57 seconds to deallocate network for instance. [ 1165.088687] env[62692]: DEBUG nova.network.neutron [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1165.097289] env[62692]: DEBUG oslo_concurrency.lockutils [req-cf126989-e0b8-407d-a75d-6ce3b0c4128a req-ef0707dc-85e3-4657-af93-f81ce84e302b service nova] Releasing lock "refresh_cache-973ac4ab-5f82-4108-8076-9a370ca9d914" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1165.145871] env[62692]: DEBUG oslo_vmware.api [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141984, 'name': RelocateVM_Task} progress is 35%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.306315] env[62692]: DEBUG nova.network.neutron [None req-4c5dba13-620e-4a18-a830-3de73c7e81c6 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Updating instance_info_cache with network_info: [{"id": "144e96aa-d02a-48e4-a91e-73909db5c202", "address": "fa:16:3e:23:7e:07", "network": {"id": "50c1330e-1a5f-42d1-9987-79ddcc0ad7aa", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-813901268-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b1c44eb72b334ee3add29026d312e37b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f92f0b92-d6fb-4d00-8ad5-6b3809ed5493", "external-id": "nsx-vlan-transportzone-48", "segmentation_id": 48, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap144e96aa-d0", "ovs_interfaceid": "144e96aa-d02a-48e4-a91e-73909db5c202", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1165.312094] env[62692]: DEBUG nova.network.neutron [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Updating instance_info_cache with network_info: [{"id": "d463b030-eded-4099-ae6f-2aea2d1af078", "address": "fa:16:3e:db:5f:6a", "network": {"id": "189f2464-2a0a-4637-b0f3-2f6ff8fdb324", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1462324860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f94f7500bf524e3890734a58937b80c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd463b030-ed", "ovs_interfaceid": "d463b030-eded-4099-ae6f-2aea2d1af078", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1165.416517] env[62692]: DEBUG oslo_vmware.api [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141985, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.500512] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.500785] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.501245] env[62692]: DEBUG nova.objects.instance [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lazy-loading 'resources' on Instance uuid cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1165.650469] env[62692]: DEBUG oslo_vmware.api [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141984, 'name': RelocateVM_Task} progress is 49%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.778942] env[62692]: DEBUG nova.compute.manager [req-3c55851e-12a6-4fa2-8933-441966263c47 req-f0d52ad6-161d-4465-adb3-0f179ac529a0 service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Received event network-vif-plugged-d463b030-eded-4099-ae6f-2aea2d1af078 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1165.779226] env[62692]: DEBUG oslo_concurrency.lockutils [req-3c55851e-12a6-4fa2-8933-441966263c47 req-f0d52ad6-161d-4465-adb3-0f179ac529a0 service nova] Acquiring lock "8abacc6f-054f-4c4f-abb2-83f20748cf1c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.779449] env[62692]: DEBUG oslo_concurrency.lockutils [req-3c55851e-12a6-4fa2-8933-441966263c47 req-f0d52ad6-161d-4465-adb3-0f179ac529a0 service nova] Lock "8abacc6f-054f-4c4f-abb2-83f20748cf1c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.779624] env[62692]: DEBUG oslo_concurrency.lockutils [req-3c55851e-12a6-4fa2-8933-441966263c47 req-f0d52ad6-161d-4465-adb3-0f179ac529a0 service nova] Lock "8abacc6f-054f-4c4f-abb2-83f20748cf1c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.779844] env[62692]: DEBUG nova.compute.manager [req-3c55851e-12a6-4fa2-8933-441966263c47 req-f0d52ad6-161d-4465-adb3-0f179ac529a0 service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] No waiting events found dispatching network-vif-plugged-d463b030-eded-4099-ae6f-2aea2d1af078 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1165.779972] env[62692]: WARNING nova.compute.manager [req-3c55851e-12a6-4fa2-8933-441966263c47 req-f0d52ad6-161d-4465-adb3-0f179ac529a0 service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Received unexpected event network-vif-plugged-d463b030-eded-4099-ae6f-2aea2d1af078 for instance with vm_state building and task_state spawning. [ 1165.780194] env[62692]: DEBUG nova.compute.manager [req-3c55851e-12a6-4fa2-8933-441966263c47 req-f0d52ad6-161d-4465-adb3-0f179ac529a0 service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Received event network-changed-d463b030-eded-4099-ae6f-2aea2d1af078 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1165.780439] env[62692]: DEBUG nova.compute.manager [req-3c55851e-12a6-4fa2-8933-441966263c47 req-f0d52ad6-161d-4465-adb3-0f179ac529a0 service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Refreshing instance network info cache due to event network-changed-d463b030-eded-4099-ae6f-2aea2d1af078. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1165.780533] env[62692]: DEBUG oslo_concurrency.lockutils [req-3c55851e-12a6-4fa2-8933-441966263c47 req-f0d52ad6-161d-4465-adb3-0f179ac529a0 service nova] Acquiring lock "refresh_cache-8abacc6f-054f-4c4f-abb2-83f20748cf1c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1165.809405] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4c5dba13-620e-4a18-a830-3de73c7e81c6 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Releasing lock "refresh_cache-2b21e8a1-5f5e-4f1f-bae6-512f1213eb48" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1165.810387] env[62692]: DEBUG nova.objects.instance [None req-4c5dba13-620e-4a18-a830-3de73c7e81c6 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Lazy-loading 'flavor' on Instance uuid 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1165.814365] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Releasing lock "refresh_cache-8abacc6f-054f-4c4f-abb2-83f20748cf1c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1165.814754] env[62692]: DEBUG nova.compute.manager [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Instance network_info: |[{"id": "d463b030-eded-4099-ae6f-2aea2d1af078", "address": "fa:16:3e:db:5f:6a", "network": {"id": "189f2464-2a0a-4637-b0f3-2f6ff8fdb324", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1462324860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f94f7500bf524e3890734a58937b80c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd463b030-ed", "ovs_interfaceid": "d463b030-eded-4099-ae6f-2aea2d1af078", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1165.815102] env[62692]: DEBUG oslo_concurrency.lockutils [req-3c55851e-12a6-4fa2-8933-441966263c47 req-f0d52ad6-161d-4465-adb3-0f179ac529a0 service nova] Acquired lock "refresh_cache-8abacc6f-054f-4c4f-abb2-83f20748cf1c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1165.815307] env[62692]: DEBUG nova.network.neutron [req-3c55851e-12a6-4fa2-8933-441966263c47 req-f0d52ad6-161d-4465-adb3-0f179ac529a0 service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Refreshing network info cache for port d463b030-eded-4099-ae6f-2aea2d1af078 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1165.817141] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:db:5f:6a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '669e4919-e0ad-4e23-9f23-4c5f2be0d858', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd463b030-eded-4099-ae6f-2aea2d1af078', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1165.824565] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Creating folder: Project (f94f7500bf524e3890734a58937b80c4). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1165.828085] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2474ac31-e63e-4ca3-865d-31854071b907 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.841661] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Created folder: Project (f94f7500bf524e3890734a58937b80c4) in parent group-v248868. [ 1165.841887] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Creating folder: Instances. Parent ref: group-v249077. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1165.842231] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2d552a1e-5fb7-455c-9d57-a837bebebebf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.853710] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Created folder: Instances in parent group-v249077. [ 1165.854392] env[62692]: DEBUG oslo.service.loopingcall [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1165.854392] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1165.854590] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fa631b7b-5cdc-4c60-94b3-6dc24dcc3ffb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.878163] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1165.878163] env[62692]: value = "task-1141988" [ 1165.878163] env[62692]: _type = "Task" [ 1165.878163] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.887705] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141988, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.913018] env[62692]: DEBUG oslo_vmware.api [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141985, 'name': PowerOffVM_Task, 'duration_secs': 0.642059} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.913361] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1165.913539] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1165.914076] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-27b38e4f-7fc7-4ed6-a758-3052d6645a3f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.042266] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1166.042536] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1166.042730] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Deleting the datastore file [datastore2] 94a07505-661f-4595-9024-8a9da9f2c2fc {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1166.043107] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cf6fbd34-a951-4fd5-8d4e-41d39260ab2c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.051228] env[62692]: DEBUG oslo_vmware.api [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for the task: (returnval){ [ 1166.051228] env[62692]: value = "task-1141990" [ 1166.051228] env[62692]: _type = "Task" [ 1166.051228] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.060989] env[62692]: DEBUG oslo_vmware.api [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141990, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.107311] env[62692]: DEBUG nova.network.neutron [req-3c55851e-12a6-4fa2-8933-441966263c47 req-f0d52ad6-161d-4465-adb3-0f179ac529a0 service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Updated VIF entry in instance network info cache for port d463b030-eded-4099-ae6f-2aea2d1af078. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1166.107732] env[62692]: DEBUG nova.network.neutron [req-3c55851e-12a6-4fa2-8933-441966263c47 req-f0d52ad6-161d-4465-adb3-0f179ac529a0 service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Updating instance_info_cache with network_info: [{"id": "d463b030-eded-4099-ae6f-2aea2d1af078", "address": "fa:16:3e:db:5f:6a", "network": {"id": "189f2464-2a0a-4637-b0f3-2f6ff8fdb324", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1462324860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f94f7500bf524e3890734a58937b80c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd463b030-ed", "ovs_interfaceid": "d463b030-eded-4099-ae6f-2aea2d1af078", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1166.150819] env[62692]: DEBUG oslo_vmware.api [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141984, 'name': RelocateVM_Task} progress is 62%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.287214] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8440f5a-acd4-4791-ab3f-82329943f812 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.296992] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a03a6c5-cdc1-484b-bbb7-def1f11fbfa4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.333734] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d12c4cac-55db-4ef3-81f1-f72fd176d12c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.337378] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9776b670-d61e-43b0-8814-6ee4717a2365 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.363421] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2976b9e-593b-4a31-b335-8834680a23a0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.368023] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c5dba13-620e-4a18-a830-3de73c7e81c6 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1166.368023] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-09a5ebe9-eda2-4c50-b47e-95644f2b1150 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.383035] env[62692]: DEBUG nova.compute.provider_tree [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1166.390176] env[62692]: DEBUG oslo_vmware.api [None req-4c5dba13-620e-4a18-a830-3de73c7e81c6 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1166.390176] env[62692]: value = "task-1141991" [ 1166.390176] env[62692]: _type = "Task" [ 1166.390176] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.398459] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1141988, 'name': CreateVM_Task, 'duration_secs': 0.401249} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.399849] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1166.400680] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1166.400852] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1166.401228] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1166.402021] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b68094a3-aff7-45ec-9a3d-c688849d2557 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.407464] env[62692]: DEBUG oslo_vmware.api [None req-4c5dba13-620e-4a18-a830-3de73c7e81c6 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141991, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.412558] env[62692]: DEBUG oslo_vmware.api [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1166.412558] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5238d3a2-9314-3fc7-b959-c3b8fdabc96d" [ 1166.412558] env[62692]: _type = "Task" [ 1166.412558] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.423170] env[62692]: DEBUG oslo_vmware.api [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5238d3a2-9314-3fc7-b959-c3b8fdabc96d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.562205] env[62692]: DEBUG oslo_vmware.api [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141990, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.611276] env[62692]: DEBUG oslo_concurrency.lockutils [req-3c55851e-12a6-4fa2-8933-441966263c47 req-f0d52ad6-161d-4465-adb3-0f179ac529a0 service nova] Releasing lock "refresh_cache-8abacc6f-054f-4c4f-abb2-83f20748cf1c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1166.648593] env[62692]: DEBUG oslo_vmware.api [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141984, 'name': RelocateVM_Task} progress is 75%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.693950] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1166.693950] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1166.890762] env[62692]: DEBUG nova.scheduler.client.report [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1166.904747] env[62692]: DEBUG oslo_vmware.api [None req-4c5dba13-620e-4a18-a830-3de73c7e81c6 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141991, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.924303] env[62692]: DEBUG oslo_vmware.api [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5238d3a2-9314-3fc7-b959-c3b8fdabc96d, 'name': SearchDatastore_Task, 'duration_secs': 0.030108} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.924639] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1166.924965] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1166.925250] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1166.925412] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1166.925595] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1166.926276] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a19436c7-295d-43fe-adb1-62c04fde27f2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.936446] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1166.936808] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1166.937527] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-182f9243-c914-489b-ae18-8739ab5a7215 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.944480] env[62692]: DEBUG oslo_vmware.api [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1166.944480] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]526f854e-f357-7e24-1aa8-f9f205b4f727" [ 1166.944480] env[62692]: _type = "Task" [ 1166.944480] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.953895] env[62692]: DEBUG oslo_vmware.api [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]526f854e-f357-7e24-1aa8-f9f205b4f727, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.063415] env[62692]: DEBUG oslo_vmware.api [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141990, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.149231] env[62692]: DEBUG oslo_vmware.api [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141984, 'name': RelocateVM_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.200740] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1167.200740] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Starting heal instance info cache {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1167.399130] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.898s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1167.407758] env[62692]: DEBUG oslo_vmware.api [None req-4c5dba13-620e-4a18-a830-3de73c7e81c6 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141991, 'name': PowerOffVM_Task, 'duration_secs': 0.989881} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.408091] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c5dba13-620e-4a18-a830-3de73c7e81c6 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1167.413985] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c5dba13-620e-4a18-a830-3de73c7e81c6 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Reconfiguring VM instance instance-0000006d to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1167.414349] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-70f112d2-9497-4f25-a2f3-44ab9405a6c6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.430234] env[62692]: INFO nova.scheduler.client.report [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Deleted allocations for instance cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43 [ 1167.437872] env[62692]: DEBUG oslo_vmware.api [None req-4c5dba13-620e-4a18-a830-3de73c7e81c6 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1167.437872] env[62692]: value = "task-1141992" [ 1167.437872] env[62692]: _type = "Task" [ 1167.437872] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.447030] env[62692]: DEBUG oslo_vmware.api [None req-4c5dba13-620e-4a18-a830-3de73c7e81c6 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141992, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.456373] env[62692]: DEBUG oslo_vmware.api [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]526f854e-f357-7e24-1aa8-f9f205b4f727, 'name': SearchDatastore_Task, 'duration_secs': 0.024835} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.457315] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68c93a58-0ee5-4642-bd4b-db9f6089ffea {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.463520] env[62692]: DEBUG oslo_vmware.api [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1167.463520] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]529a017b-31a9-a8d7-b53b-0131480fdd42" [ 1167.463520] env[62692]: _type = "Task" [ 1167.463520] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.474831] env[62692]: DEBUG oslo_vmware.api [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]529a017b-31a9-a8d7-b53b-0131480fdd42, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.564562] env[62692]: DEBUG oslo_vmware.api [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Task: {'id': task-1141990, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.321664} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.564872] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1167.565191] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1167.566037] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1167.566037] env[62692]: INFO nova.compute.manager [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Took 2.68 seconds to destroy the instance on the hypervisor. [ 1167.566037] env[62692]: DEBUG oslo.service.loopingcall [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1167.566411] env[62692]: DEBUG nova.compute.manager [-] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1167.566523] env[62692]: DEBUG nova.network.neutron [-] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1167.649588] env[62692]: DEBUG oslo_vmware.api [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141984, 'name': RelocateVM_Task} progress is 97%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.704577] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "refresh_cache-90932f05-fad3-4c6d-87ae-ab059351b0be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1167.704734] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquired lock "refresh_cache-90932f05-fad3-4c6d-87ae-ab059351b0be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1167.704884] env[62692]: DEBUG nova.network.neutron [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Forcefully refreshing network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1167.939158] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1c00205a-d2e3-4496-b7c0-bb4bc38fa86a tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.656s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1167.951480] env[62692]: DEBUG oslo_vmware.api [None req-4c5dba13-620e-4a18-a830-3de73c7e81c6 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141992, 'name': ReconfigVM_Task, 'duration_secs': 0.382681} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.951480] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c5dba13-620e-4a18-a830-3de73c7e81c6 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Reconfigured VM instance instance-0000006d to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1167.951692] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c5dba13-620e-4a18-a830-3de73c7e81c6 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1167.952520] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4ac9e069-4b5e-456a-9f6d-69484052d589 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.961986] env[62692]: DEBUG oslo_vmware.api [None req-4c5dba13-620e-4a18-a830-3de73c7e81c6 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1167.961986] env[62692]: value = "task-1141993" [ 1167.961986] env[62692]: _type = "Task" [ 1167.961986] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.980447] env[62692]: DEBUG oslo_vmware.api [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]529a017b-31a9-a8d7-b53b-0131480fdd42, 'name': SearchDatastore_Task, 'duration_secs': 0.028902} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.981081] env[62692]: DEBUG oslo_vmware.api [None req-4c5dba13-620e-4a18-a830-3de73c7e81c6 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141993, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.981459] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1167.981661] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 8abacc6f-054f-4c4f-abb2-83f20748cf1c/8abacc6f-054f-4c4f-abb2-83f20748cf1c.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1167.981972] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0bb9a2df-1700-4ce1-988e-feeda4b70a2e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.987227] env[62692]: DEBUG nova.compute.manager [req-5a782fec-171d-4f79-87ca-e044b253ba8e req-18a54a68-ddb7-4bae-bb44-85bd1a935a2d service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Received event network-vif-deleted-d51ffd3f-c023-45e2-8bfa-1424b7ccb188 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1167.987503] env[62692]: INFO nova.compute.manager [req-5a782fec-171d-4f79-87ca-e044b253ba8e req-18a54a68-ddb7-4bae-bb44-85bd1a935a2d service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Neutron deleted interface d51ffd3f-c023-45e2-8bfa-1424b7ccb188; detaching it from the instance and deleting it from the info cache [ 1167.987714] env[62692]: DEBUG nova.network.neutron [req-5a782fec-171d-4f79-87ca-e044b253ba8e req-18a54a68-ddb7-4bae-bb44-85bd1a935a2d service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1167.994855] env[62692]: DEBUG oslo_vmware.api [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1167.994855] env[62692]: value = "task-1141994" [ 1167.994855] env[62692]: _type = "Task" [ 1167.994855] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.005485] env[62692]: DEBUG oslo_vmware.api [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1141994, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.150939] env[62692]: DEBUG oslo_vmware.api [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141984, 'name': RelocateVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.427863] env[62692]: DEBUG nova.network.neutron [-] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1168.474108] env[62692]: DEBUG oslo_vmware.api [None req-4c5dba13-620e-4a18-a830-3de73c7e81c6 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141993, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.491074] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9f437bae-884a-414b-ab9e-e2cb3e923b4b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.507777] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2ef7493-6fd5-409f-90f7-6ab6603f9be4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.524181] env[62692]: DEBUG oslo_vmware.api [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1141994, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.548652] env[62692]: DEBUG nova.compute.manager [req-5a782fec-171d-4f79-87ca-e044b253ba8e req-18a54a68-ddb7-4bae-bb44-85bd1a935a2d service nova] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Detach interface failed, port_id=d51ffd3f-c023-45e2-8bfa-1424b7ccb188, reason: Instance 94a07505-661f-4595-9024-8a9da9f2c2fc could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1168.650912] env[62692]: DEBUG oslo_vmware.api [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141984, 'name': RelocateVM_Task, 'duration_secs': 3.524373} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.652531] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Volume attach. Driver type: vmdk {{(pid=62692) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1168.652531] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249070', 'volume_id': 'd3635284-5881-414b-8d43-075170b6ad32', 'name': 'volume-d3635284-5881-414b-8d43-075170b6ad32', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '973ac4ab-5f82-4108-8076-9a370ca9d914', 'attached_at': '', 'detached_at': '', 'volume_id': 'd3635284-5881-414b-8d43-075170b6ad32', 'serial': 'd3635284-5881-414b-8d43-075170b6ad32'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1168.652531] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-601df439-e6ea-413c-8be9-9c1d40847627 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.669370] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b10c8cb-bbf2-4d87-9d9c-909ad8ae7344 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.695453] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] volume-d3635284-5881-414b-8d43-075170b6ad32/volume-d3635284-5881-414b-8d43-075170b6ad32.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1168.696040] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1cc9ade2-3f24-4d30-9e32-3f49d7306010 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.719479] env[62692]: DEBUG oslo_vmware.api [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1168.719479] env[62692]: value = "task-1141995" [ 1168.719479] env[62692]: _type = "Task" [ 1168.719479] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.729028] env[62692]: DEBUG oslo_vmware.api [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141995, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.734839] env[62692]: DEBUG oslo_concurrency.lockutils [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "9d3c0909-6207-46b7-b83b-de4d7d626fd7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1168.735192] env[62692]: DEBUG oslo_concurrency.lockutils [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "9d3c0909-6207-46b7-b83b-de4d7d626fd7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.813808] env[62692]: DEBUG oslo_vmware.rw_handles [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525e3a00-51a7-6640-0ddd-1f99f5092c6f/disk-0.vmdk. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1168.816165] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8428ade-3009-466b-a2a6-a19e7d8925bf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.822818] env[62692]: DEBUG oslo_vmware.rw_handles [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525e3a00-51a7-6640-0ddd-1f99f5092c6f/disk-0.vmdk is in state: ready. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1168.823082] env[62692]: ERROR oslo_vmware.rw_handles [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525e3a00-51a7-6640-0ddd-1f99f5092c6f/disk-0.vmdk due to incomplete transfer. [ 1168.823380] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-cd1a6002-87ed-405b-93d6-814d6f2d7485 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.834527] env[62692]: DEBUG oslo_vmware.rw_handles [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525e3a00-51a7-6640-0ddd-1f99f5092c6f/disk-0.vmdk. {{(pid=62692) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1168.834527] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Uploaded image a661e426-af79-4233-8845-0c9ddc6538a5 to the Glance image server {{(pid=62692) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1168.836438] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Destroying the VM {{(pid=62692) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1168.836729] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-3671d96b-fc7a-4189-8b37-6c4fd5f6012f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.844617] env[62692]: DEBUG oslo_vmware.api [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1168.844617] env[62692]: value = "task-1141996" [ 1168.844617] env[62692]: _type = "Task" [ 1168.844617] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.859468] env[62692]: DEBUG oslo_vmware.api [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141996, 'name': Destroy_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.932981] env[62692]: INFO nova.compute.manager [-] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Took 1.37 seconds to deallocate network for instance. [ 1168.977777] env[62692]: DEBUG oslo_vmware.api [None req-4c5dba13-620e-4a18-a830-3de73c7e81c6 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1141993, 'name': PowerOnVM_Task, 'duration_secs': 0.553524} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.978706] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c5dba13-620e-4a18-a830-3de73c7e81c6 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1168.979503] env[62692]: DEBUG nova.compute.manager [None req-4c5dba13-620e-4a18-a830-3de73c7e81c6 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1168.980769] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6691b8ab-0a6a-49e9-bd83-f4a3e5639326 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.005862] env[62692]: DEBUG oslo_vmware.api [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1141994, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.609696} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.006064] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 8abacc6f-054f-4c4f-abb2-83f20748cf1c/8abacc6f-054f-4c4f-abb2-83f20748cf1c.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1169.007043] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1169.007043] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5a3acf72-5553-4965-9e69-9ab11cadfa79 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.015140] env[62692]: DEBUG oslo_vmware.api [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1169.015140] env[62692]: value = "task-1141997" [ 1169.015140] env[62692]: _type = "Task" [ 1169.015140] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.025644] env[62692]: DEBUG oslo_vmware.api [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1141997, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.027278] env[62692]: DEBUG nova.network.neutron [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Updating instance_info_cache with network_info: [{"id": "21e76034-6c0f-4938-afda-0d005cf45759", "address": "fa:16:3e:2d:c0:4a", "network": {"id": "77695f27-3027-4a92-8b9b-0a90cdc53f62", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-337681783-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697b4281872f4547ad3d051e644b95ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21e76034-6c", "ovs_interfaceid": "21e76034-6c0f-4938-afda-0d005cf45759", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1169.230393] env[62692]: DEBUG oslo_vmware.api [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141995, 'name': ReconfigVM_Task, 'duration_secs': 0.342297} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.230520] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Reconfigured VM instance instance-00000070 to attach disk [datastore1] volume-d3635284-5881-414b-8d43-075170b6ad32/volume-d3635284-5881-414b-8d43-075170b6ad32.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1169.235219] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-385c93e9-4c97-4f50-ad3f-1ef2a5ebb408 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.245856] env[62692]: DEBUG nova.compute.manager [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1169.256249] env[62692]: DEBUG oslo_vmware.api [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1169.256249] env[62692]: value = "task-1141998" [ 1169.256249] env[62692]: _type = "Task" [ 1169.256249] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.264914] env[62692]: DEBUG oslo_vmware.api [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141998, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.357706] env[62692]: DEBUG oslo_vmware.api [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141996, 'name': Destroy_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.441587] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.441880] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.442207] env[62692]: DEBUG nova.objects.instance [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lazy-loading 'resources' on Instance uuid 94a07505-661f-4595-9024-8a9da9f2c2fc {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1169.526753] env[62692]: DEBUG oslo_vmware.api [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1141997, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076753} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.526753] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1169.528265] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50ca56b8-5003-4668-b497-9906bb105c13 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.530229] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Releasing lock "refresh_cache-90932f05-fad3-4c6d-87ae-ab059351b0be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1169.530485] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Updated the network info_cache for instance {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1169.530774] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1169.531596] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1169.531877] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1169.532163] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1169.532366] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1169.532590] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1169.532741] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62692) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1169.532899] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager.update_available_resource {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1169.552633] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Reconfiguring VM instance instance-00000071 to attach disk [datastore1] 8abacc6f-054f-4c4f-abb2-83f20748cf1c/8abacc6f-054f-4c4f-abb2-83f20748cf1c.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1169.553749] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.554283] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff0597ab-3816-488f-8d32-8f9e217b1823 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.575646] env[62692]: DEBUG oslo_vmware.api [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1169.575646] env[62692]: value = "task-1141999" [ 1169.575646] env[62692]: _type = "Task" [ 1169.575646] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.584668] env[62692]: DEBUG oslo_vmware.api [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1141999, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.768713] env[62692]: DEBUG oslo_vmware.api [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1141998, 'name': ReconfigVM_Task, 'duration_secs': 0.139387} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.769062] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249070', 'volume_id': 'd3635284-5881-414b-8d43-075170b6ad32', 'name': 'volume-d3635284-5881-414b-8d43-075170b6ad32', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '973ac4ab-5f82-4108-8076-9a370ca9d914', 'attached_at': '', 'detached_at': '', 'volume_id': 'd3635284-5881-414b-8d43-075170b6ad32', 'serial': 'd3635284-5881-414b-8d43-075170b6ad32'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1169.769671] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9adab400-af35-4d5c-aeb1-2b76979d76f2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.779504] env[62692]: DEBUG oslo_vmware.api [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1169.779504] env[62692]: value = "task-1142000" [ 1169.779504] env[62692]: _type = "Task" [ 1169.779504] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.781460] env[62692]: DEBUG oslo_concurrency.lockutils [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.810693] env[62692]: DEBUG oslo_vmware.api [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1142000, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.857188] env[62692]: DEBUG oslo_vmware.api [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1141996, 'name': Destroy_Task, 'duration_secs': 0.885953} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.857188] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Destroyed the VM [ 1169.857547] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Deleting Snapshot of the VM instance {{(pid=62692) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1169.857766] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-28b790d3-0d38-41b4-8e65-71c8864ebcca {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.868092] env[62692]: DEBUG oslo_vmware.api [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1169.868092] env[62692]: value = "task-1142001" [ 1169.868092] env[62692]: _type = "Task" [ 1169.868092] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.883790] env[62692]: DEBUG oslo_vmware.api [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142001, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.087903] env[62692]: DEBUG oslo_vmware.api [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1141999, 'name': ReconfigVM_Task, 'duration_secs': 0.312148} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.090618] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Reconfigured VM instance instance-00000071 to attach disk [datastore1] 8abacc6f-054f-4c4f-abb2-83f20748cf1c/8abacc6f-054f-4c4f-abb2-83f20748cf1c.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1170.092524] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-95e6081c-9fe3-4817-b2f3-2f64bbf12e11 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.101531] env[62692]: DEBUG oslo_vmware.api [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1170.101531] env[62692]: value = "task-1142002" [ 1170.101531] env[62692]: _type = "Task" [ 1170.101531] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.112829] env[62692]: DEBUG oslo_vmware.api [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142002, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.133743] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1140bc8e-479d-4389-9306-702b81d4a64a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.142193] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f808814f-b801-4527-b301-c794b116704d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.184636] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fc3fb4f-f8b8-4f1a-8345-d376978ea5a9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.193775] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeaafcb1-459c-4f06-8d51-b5dfe752ebe8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.210865] env[62692]: DEBUG nova.compute.provider_tree [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1170.295587] env[62692]: DEBUG oslo_vmware.api [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1142000, 'name': Rename_Task, 'duration_secs': 0.167111} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.295587] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1170.295587] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-03a55415-e716-40f3-819b-0a2713d1c115 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.303144] env[62692]: DEBUG oslo_vmware.api [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1170.303144] env[62692]: value = "task-1142003" [ 1170.303144] env[62692]: _type = "Task" [ 1170.303144] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.311892] env[62692]: DEBUG oslo_vmware.api [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1142003, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.379988] env[62692]: DEBUG oslo_vmware.api [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142001, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.615354] env[62692]: DEBUG oslo_vmware.api [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142002, 'name': Rename_Task, 'duration_secs': 0.163506} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.615634] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1170.615889] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-334b4088-5669-42c9-82a6-2d64ab9f01d4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.623930] env[62692]: DEBUG oslo_vmware.api [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1170.623930] env[62692]: value = "task-1142004" [ 1170.623930] env[62692]: _type = "Task" [ 1170.623930] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.635851] env[62692]: DEBUG oslo_vmware.api [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142004, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.651585] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquiring lock "2b21e8a1-5f5e-4f1f-bae6-512f1213eb48" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1170.652188] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Lock "2b21e8a1-5f5e-4f1f-bae6-512f1213eb48" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.652465] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquiring lock "2b21e8a1-5f5e-4f1f-bae6-512f1213eb48-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1170.652687] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Lock "2b21e8a1-5f5e-4f1f-bae6-512f1213eb48-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.652867] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Lock "2b21e8a1-5f5e-4f1f-bae6-512f1213eb48-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.655394] env[62692]: INFO nova.compute.manager [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Terminating instance [ 1170.657286] env[62692]: DEBUG nova.compute.manager [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1170.657491] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1170.658331] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96e7e429-2048-42dc-a57b-e9dab765128a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.666471] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1170.666767] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-25f72a4a-1b59-4f0b-b30c-9add07315c87 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.675451] env[62692]: DEBUG oslo_vmware.api [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1170.675451] env[62692]: value = "task-1142005" [ 1170.675451] env[62692]: _type = "Task" [ 1170.675451] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.687091] env[62692]: DEBUG oslo_vmware.api [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1142005, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.715629] env[62692]: DEBUG nova.scheduler.client.report [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1170.816562] env[62692]: DEBUG oslo_vmware.api [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1142003, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.878877] env[62692]: DEBUG oslo_vmware.api [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142001, 'name': RemoveSnapshot_Task, 'duration_secs': 0.946099} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.879188] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Deleted Snapshot of the VM instance {{(pid=62692) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1170.879431] env[62692]: INFO nova.compute.manager [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Took 18.85 seconds to snapshot the instance on the hypervisor. [ 1171.045359] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "c2f2a485-363e-41af-9b3f-37092947ca08" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1171.045535] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "c2f2a485-363e-41af-9b3f-37092947ca08" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1171.047363] env[62692]: INFO nova.compute.manager [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Shelving [ 1171.136722] env[62692]: DEBUG oslo_vmware.api [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142004, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.187823] env[62692]: DEBUG oslo_vmware.api [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1142005, 'name': PowerOffVM_Task, 'duration_secs': 0.216378} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.188144] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1171.188331] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1171.188600] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bbd7437f-c5f7-4850-9e32-1808ecb5ba12 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.220422] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.778s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1171.223428] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.669s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1171.223808] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1171.223987] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62692) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1171.224340] env[62692]: DEBUG oslo_concurrency.lockutils [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.443s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1171.226324] env[62692]: INFO nova.compute.claims [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1171.232237] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0411475-e206-4736-8a9f-388af8e78b1c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.239938] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e94ac24b-ca9d-4eec-8754-5e508f8ce1b1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.263391] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00288108-6e06-41d2-9a1d-139481da646c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.266266] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1171.266530] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1171.266744] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Deleting the datastore file [datastore2] 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1171.268636] env[62692]: INFO nova.scheduler.client.report [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Deleted allocations for instance 94a07505-661f-4595-9024-8a9da9f2c2fc [ 1171.269748] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9ce46a22-6438-4f5d-8939-19c58e7cf6d3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.280260] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-866a429b-5ad2-4d92-9c16-abc8baf1d846 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.283728] env[62692]: DEBUG oslo_vmware.api [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1171.283728] env[62692]: value = "task-1142007" [ 1171.283728] env[62692]: _type = "Task" [ 1171.283728] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.319333] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180207MB free_disk=166GB free_vcpus=48 pci_devices=None {{(pid=62692) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1171.319578] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1171.328937] env[62692]: DEBUG oslo_vmware.api [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1142007, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.335281] env[62692]: DEBUG oslo_vmware.api [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1142003, 'name': PowerOnVM_Task, 'duration_secs': 0.519103} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.335577] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1171.335811] env[62692]: INFO nova.compute.manager [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Took 8.50 seconds to spawn the instance on the hypervisor. [ 1171.336020] env[62692]: DEBUG nova.compute.manager [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1171.336844] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1a6baf4-3ac2-4875-a044-d14a14782443 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.432286] env[62692]: DEBUG nova.compute.manager [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Found 3 images (rotation: 2) {{(pid=62692) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1171.432531] env[62692]: DEBUG nova.compute.manager [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Rotating out 1 backups {{(pid=62692) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4562}} [ 1171.432808] env[62692]: DEBUG nova.compute.manager [None req-333ba9ef-fa64-4ea7-8de5-a268464bd65b tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Deleting image b7b0878a-b2c2-4474-9dbb-47b954aa8a86 {{(pid=62692) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4567}} [ 1171.552974] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1171.553281] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-500cfee4-bec2-4a19-b812-c40b204f0abd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.561846] env[62692]: DEBUG oslo_vmware.api [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1171.561846] env[62692]: value = "task-1142008" [ 1171.561846] env[62692]: _type = "Task" [ 1171.561846] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.573862] env[62692]: DEBUG oslo_vmware.api [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142008, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.635503] env[62692]: DEBUG oslo_vmware.api [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142004, 'name': PowerOnVM_Task, 'duration_secs': 0.816113} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.635768] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1171.635975] env[62692]: INFO nova.compute.manager [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Took 8.18 seconds to spawn the instance on the hypervisor. [ 1171.636274] env[62692]: DEBUG nova.compute.manager [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1171.637090] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f4dfe4b-3499-4bc6-86d5-f80007313532 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.777938] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c3451ce0-7745-430c-9a44-418091cb5063 tempest-AttachInterfacesTestJSON-369176474 tempest-AttachInterfacesTestJSON-369176474-project-member] Lock "94a07505-661f-4595-9024-8a9da9f2c2fc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.899s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1171.809144] env[62692]: DEBUG oslo_vmware.api [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1142007, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154447} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.809493] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1171.809703] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1171.809910] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1171.810100] env[62692]: INFO nova.compute.manager [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1171.812042] env[62692]: DEBUG oslo.service.loopingcall [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1171.812311] env[62692]: DEBUG nova.compute.manager [-] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1171.812457] env[62692]: DEBUG nova.network.neutron [-] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1171.854185] env[62692]: INFO nova.compute.manager [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Took 15.83 seconds to build instance. [ 1172.074482] env[62692]: DEBUG oslo_vmware.api [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142008, 'name': PowerOffVM_Task, 'duration_secs': 0.420455} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.074938] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1172.076018] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dac0e091-540b-4464-9df0-19fc4f56ea76 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.097088] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ae8ea50-e81e-4d92-ba6c-4bf1cb8070cd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.156410] env[62692]: INFO nova.compute.manager [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Took 14.85 seconds to build instance. [ 1172.173948] env[62692]: DEBUG nova.compute.manager [req-bb15fc83-9919-4556-a5d8-ae7b6d0fd216 req-4b48e940-0fff-485c-bf0d-d5c2e47c5d6f service nova] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Received event network-vif-deleted-144e96aa-d02a-48e4-a91e-73909db5c202 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1172.173948] env[62692]: INFO nova.compute.manager [req-bb15fc83-9919-4556-a5d8-ae7b6d0fd216 req-4b48e940-0fff-485c-bf0d-d5c2e47c5d6f service nova] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Neutron deleted interface 144e96aa-d02a-48e4-a91e-73909db5c202; detaching it from the instance and deleting it from the info cache [ 1172.173948] env[62692]: DEBUG nova.network.neutron [req-bb15fc83-9919-4556-a5d8-ae7b6d0fd216 req-4b48e940-0fff-485c-bf0d-d5c2e47c5d6f service nova] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1172.357947] env[62692]: DEBUG oslo_concurrency.lockutils [None req-edb11488-4394-402c-929b-12f0bb47c7df tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "973ac4ab-5f82-4108-8076-9a370ca9d914" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.348s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1172.453309] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46feb23b-8889-4c72-a7e7-c02b3e59fa49 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.466558] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7785de5-47f8-4de9-8a59-0137676146c4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.507838] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddd0a3f0-2454-440e-a97c-ebb9d1481d35 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.517317] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ba636fd-e892-418d-8826-a840a18b3ce3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.535135] env[62692]: DEBUG nova.compute.provider_tree [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1172.613793] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Creating Snapshot of the VM instance {{(pid=62692) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1172.614148] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-92d7557e-b2a6-4e0a-880d-f6cfb2412d1f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.623601] env[62692]: DEBUG oslo_vmware.api [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1172.623601] env[62692]: value = "task-1142009" [ 1172.623601] env[62692]: _type = "Task" [ 1172.623601] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.637486] env[62692]: DEBUG nova.network.neutron [-] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1172.640089] env[62692]: DEBUG oslo_vmware.api [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142009, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.658544] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8e84dfe8-3074-4dd6-bcb2-d30efb9649ec tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "8abacc6f-054f-4c4f-abb2-83f20748cf1c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.366s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1172.679929] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c27bcb59-8340-4011-97ed-7f300de4b065 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.689049] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16bb6a14-c88f-424f-8f40-01b5f0d900af {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.729008] env[62692]: DEBUG nova.compute.manager [req-bb15fc83-9919-4556-a5d8-ae7b6d0fd216 req-4b48e940-0fff-485c-bf0d-d5c2e47c5d6f service nova] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Detach interface failed, port_id=144e96aa-d02a-48e4-a91e-73909db5c202, reason: Instance 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1173.001490] env[62692]: DEBUG oslo_concurrency.lockutils [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquiring lock "9d754b1a-0e1b-4200-804e-9e4fa3bd06bf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1173.001733] env[62692]: DEBUG oslo_concurrency.lockutils [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "9d754b1a-0e1b-4200-804e-9e4fa3bd06bf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1173.040236] env[62692]: DEBUG nova.scheduler.client.report [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1173.136819] env[62692]: DEBUG oslo_vmware.api [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142009, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.142128] env[62692]: INFO nova.compute.manager [-] [instance: 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48] Took 1.33 seconds to deallocate network for instance. [ 1173.504314] env[62692]: DEBUG nova.compute.manager [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1173.548372] env[62692]: DEBUG oslo_concurrency.lockutils [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.324s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1173.548921] env[62692]: DEBUG nova.compute.manager [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1173.551818] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 2.232s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1173.639911] env[62692]: DEBUG oslo_vmware.api [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142009, 'name': CreateSnapshot_Task, 'duration_secs': 0.961295} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.639911] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Created Snapshot of the VM instance {{(pid=62692) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1173.640160] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29a5a25c-3035-4d0d-b568-40535406c789 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.656104] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1173.727954] env[62692]: DEBUG nova.compute.manager [req-f82588eb-d031-4d4a-ba4c-b0308144fdd4 req-64e33b4f-88ec-4000-8215-e8a52e26faea service nova] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Received event network-changed-a291ddc4-8996-4755-b347-066f38e6ec7d {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1173.727954] env[62692]: DEBUG nova.compute.manager [req-f82588eb-d031-4d4a-ba4c-b0308144fdd4 req-64e33b4f-88ec-4000-8215-e8a52e26faea service nova] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Refreshing instance network info cache due to event network-changed-a291ddc4-8996-4755-b347-066f38e6ec7d. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1173.728695] env[62692]: DEBUG oslo_concurrency.lockutils [req-f82588eb-d031-4d4a-ba4c-b0308144fdd4 req-64e33b4f-88ec-4000-8215-e8a52e26faea service nova] Acquiring lock "refresh_cache-e2383a6a-3581-40fc-a0eb-6981acdbf54a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1173.729276] env[62692]: DEBUG oslo_concurrency.lockutils [req-f82588eb-d031-4d4a-ba4c-b0308144fdd4 req-64e33b4f-88ec-4000-8215-e8a52e26faea service nova] Acquired lock "refresh_cache-e2383a6a-3581-40fc-a0eb-6981acdbf54a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1173.729276] env[62692]: DEBUG nova.network.neutron [req-f82588eb-d031-4d4a-ba4c-b0308144fdd4 req-64e33b4f-88ec-4000-8215-e8a52e26faea service nova] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Refreshing network info cache for port a291ddc4-8996-4755-b347-066f38e6ec7d {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1173.746397] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "10d13b95-e6a1-4a11-9d8c-540ce82316de" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1173.746647] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "10d13b95-e6a1-4a11-9d8c-540ce82316de" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1173.785263] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1173.785263] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1174.038344] env[62692]: DEBUG oslo_concurrency.lockutils [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1174.055511] env[62692]: DEBUG nova.compute.utils [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1174.063100] env[62692]: DEBUG nova.compute.manager [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1174.063484] env[62692]: DEBUG nova.network.neutron [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1174.133809] env[62692]: DEBUG nova.policy [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6e9cd6d33e744da4a2f678226b684845', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1037e5d436bd429391159b87a712f6e1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1174.165816] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Creating linked-clone VM from snapshot {{(pid=62692) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1174.166373] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-1e177ac0-c6a3-49dc-bca0-118ce8ab3a16 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.178167] env[62692]: DEBUG oslo_vmware.api [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1174.178167] env[62692]: value = "task-1142010" [ 1174.178167] env[62692]: _type = "Task" [ 1174.178167] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.190870] env[62692]: DEBUG oslo_vmware.api [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142010, 'name': CloneVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.249022] env[62692]: DEBUG nova.compute.manager [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1174.287779] env[62692]: DEBUG nova.compute.manager [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1174.567020] env[62692]: DEBUG nova.compute.manager [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1174.585543] env[62692]: DEBUG nova.network.neutron [req-f82588eb-d031-4d4a-ba4c-b0308144fdd4 req-64e33b4f-88ec-4000-8215-e8a52e26faea service nova] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Updated VIF entry in instance network info cache for port a291ddc4-8996-4755-b347-066f38e6ec7d. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1174.585543] env[62692]: DEBUG nova.network.neutron [req-f82588eb-d031-4d4a-ba4c-b0308144fdd4 req-64e33b4f-88ec-4000-8215-e8a52e26faea service nova] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Updating instance_info_cache with network_info: [{"id": "a291ddc4-8996-4755-b347-066f38e6ec7d", "address": "fa:16:3e:7b:10:2a", "network": {"id": "2febb660-329d-403c-8c2c-f11c0f5f547c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1806650357-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37921f67664544478cb2568def068b11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa291ddc4-89", "ovs_interfaceid": "a291ddc4-8996-4755-b347-066f38e6ec7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1174.686377] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance e2383a6a-3581-40fc-a0eb-6981acdbf54a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1174.687600] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 90932f05-fad3-4c6d-87ae-ab059351b0be actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1174.687600] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance dd8a0d53-5dc8-4b66-add7-f244e600468d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1174.687600] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance c2f2a485-363e-41af-9b3f-37092947ca08 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1174.687600] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 1c7ce408-6e9c-43bd-8d81-6acb12353ed8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1174.687600] env[62692]: WARNING nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1174.688145] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 0b7ea87c-252a-48b3-a1cd-c81e91362812 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1174.688419] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 45eccffc-f059-4bef-97da-dacb16f20f88 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1174.688809] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 8abacc6f-054f-4c4f-abb2-83f20748cf1c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1174.691095] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 9d3c0909-6207-46b7-b83b-de4d7d626fd7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1174.699162] env[62692]: DEBUG oslo_vmware.api [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142010, 'name': CloneVM_Task} progress is 94%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.708023] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3af792c1-4116-4fb2-9f15-57153c51929f tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "90932f05-fad3-4c6d-87ae-ab059351b0be" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1174.708023] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3af792c1-4116-4fb2-9f15-57153c51929f tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "90932f05-fad3-4c6d-87ae-ab059351b0be" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1174.708023] env[62692]: DEBUG nova.compute.manager [None req-3af792c1-4116-4fb2-9f15-57153c51929f tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1174.708023] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4d18857-9542-4aa1-8674-277a51e94a3d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.716025] env[62692]: DEBUG nova.compute.manager [None req-3af792c1-4116-4fb2-9f15-57153c51929f tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62692) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1174.717121] env[62692]: DEBUG nova.objects.instance [None req-3af792c1-4116-4fb2-9f15-57153c51929f tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lazy-loading 'flavor' on Instance uuid 90932f05-fad3-4c6d-87ae-ab059351b0be {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1174.762548] env[62692]: DEBUG nova.network.neutron [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Successfully created port: ea6ff473-6d45-4425-a9e9-535a999fc3a4 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1174.773206] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1174.826212] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1175.087129] env[62692]: DEBUG oslo_concurrency.lockutils [req-f82588eb-d031-4d4a-ba4c-b0308144fdd4 req-64e33b4f-88ec-4000-8215-e8a52e26faea service nova] Releasing lock "refresh_cache-e2383a6a-3581-40fc-a0eb-6981acdbf54a" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1175.191293] env[62692]: DEBUG oslo_vmware.api [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142010, 'name': CloneVM_Task} progress is 94%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.191986] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1175.229642] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3af792c1-4116-4fb2-9f15-57153c51929f tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1175.229928] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-365058ae-ef82-4d2b-b97b-73946c9aa503 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.237797] env[62692]: DEBUG oslo_vmware.api [None req-3af792c1-4116-4fb2-9f15-57153c51929f tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1175.237797] env[62692]: value = "task-1142011" [ 1175.237797] env[62692]: _type = "Task" [ 1175.237797] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.248942] env[62692]: DEBUG oslo_vmware.api [None req-3af792c1-4116-4fb2-9f15-57153c51929f tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142011, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.357343] env[62692]: DEBUG nova.compute.manager [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Stashing vm_state: active {{(pid=62692) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1175.585394] env[62692]: DEBUG nova.compute.manager [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1175.619856] env[62692]: DEBUG nova.virt.hardware [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1175.623817] env[62692]: DEBUG nova.virt.hardware [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1175.623817] env[62692]: DEBUG nova.virt.hardware [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1175.623817] env[62692]: DEBUG nova.virt.hardware [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1175.623817] env[62692]: DEBUG nova.virt.hardware [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1175.623817] env[62692]: DEBUG nova.virt.hardware [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1175.623817] env[62692]: DEBUG nova.virt.hardware [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1175.623817] env[62692]: DEBUG nova.virt.hardware [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1175.623817] env[62692]: DEBUG nova.virt.hardware [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1175.623817] env[62692]: DEBUG nova.virt.hardware [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1175.623817] env[62692]: DEBUG nova.virt.hardware [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1175.624323] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a47488d5-7294-4a2d-8834-ebc4f67eea16 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.636065] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01f69d75-b4a6-45b6-afab-37b92d521157 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.692216] env[62692]: DEBUG oslo_vmware.api [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142010, 'name': CloneVM_Task} progress is 94%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.695133] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 10d13b95-e6a1-4a11-9d8c-540ce82316de has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1175.756422] env[62692]: DEBUG oslo_vmware.api [None req-3af792c1-4116-4fb2-9f15-57153c51929f tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142011, 'name': PowerOffVM_Task, 'duration_secs': 0.213571} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.757718] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3af792c1-4116-4fb2-9f15-57153c51929f tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1175.757908] env[62692]: DEBUG nova.compute.manager [None req-3af792c1-4116-4fb2-9f15-57153c51929f tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1175.759542] env[62692]: DEBUG nova.compute.manager [req-b6d528eb-5352-4c87-98b8-415801936619 req-9832e2a9-9ae6-4297-ba17-b54c5394c6b8 service nova] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Received event network-changed-e2bdb2a3-7e2c-4329-9008-9bea98b0b80a {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1175.759542] env[62692]: DEBUG nova.compute.manager [req-b6d528eb-5352-4c87-98b8-415801936619 req-9832e2a9-9ae6-4297-ba17-b54c5394c6b8 service nova] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Refreshing instance network info cache due to event network-changed-e2bdb2a3-7e2c-4329-9008-9bea98b0b80a. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1175.759656] env[62692]: DEBUG oslo_concurrency.lockutils [req-b6d528eb-5352-4c87-98b8-415801936619 req-9832e2a9-9ae6-4297-ba17-b54c5394c6b8 service nova] Acquiring lock "refresh_cache-973ac4ab-5f82-4108-8076-9a370ca9d914" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1175.759779] env[62692]: DEBUG oslo_concurrency.lockutils [req-b6d528eb-5352-4c87-98b8-415801936619 req-9832e2a9-9ae6-4297-ba17-b54c5394c6b8 service nova] Acquired lock "refresh_cache-973ac4ab-5f82-4108-8076-9a370ca9d914" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1175.760496] env[62692]: DEBUG nova.network.neutron [req-b6d528eb-5352-4c87-98b8-415801936619 req-9832e2a9-9ae6-4297-ba17-b54c5394c6b8 service nova] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Refreshing network info cache for port e2bdb2a3-7e2c-4329-9008-9bea98b0b80a {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1175.764866] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304a57af-6950-46c7-a4c7-d5cdea414681 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.882395] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1176.191798] env[62692]: DEBUG oslo_vmware.api [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142010, 'name': CloneVM_Task} progress is 95%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.199926] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1176.255156] env[62692]: DEBUG nova.compute.manager [req-e2d9aa98-2dd5-4525-af0b-29cd9149ee14 req-1327b7e9-d115-4718-91e6-09980e47b232 service nova] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Received event network-vif-plugged-ea6ff473-6d45-4425-a9e9-535a999fc3a4 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1176.255156] env[62692]: DEBUG oslo_concurrency.lockutils [req-e2d9aa98-2dd5-4525-af0b-29cd9149ee14 req-1327b7e9-d115-4718-91e6-09980e47b232 service nova] Acquiring lock "9d3c0909-6207-46b7-b83b-de4d7d626fd7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1176.255156] env[62692]: DEBUG oslo_concurrency.lockutils [req-e2d9aa98-2dd5-4525-af0b-29cd9149ee14 req-1327b7e9-d115-4718-91e6-09980e47b232 service nova] Lock "9d3c0909-6207-46b7-b83b-de4d7d626fd7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1176.255422] env[62692]: DEBUG oslo_concurrency.lockutils [req-e2d9aa98-2dd5-4525-af0b-29cd9149ee14 req-1327b7e9-d115-4718-91e6-09980e47b232 service nova] Lock "9d3c0909-6207-46b7-b83b-de4d7d626fd7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1176.256186] env[62692]: DEBUG nova.compute.manager [req-e2d9aa98-2dd5-4525-af0b-29cd9149ee14 req-1327b7e9-d115-4718-91e6-09980e47b232 service nova] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] No waiting events found dispatching network-vif-plugged-ea6ff473-6d45-4425-a9e9-535a999fc3a4 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1176.256490] env[62692]: WARNING nova.compute.manager [req-e2d9aa98-2dd5-4525-af0b-29cd9149ee14 req-1327b7e9-d115-4718-91e6-09980e47b232 service nova] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Received unexpected event network-vif-plugged-ea6ff473-6d45-4425-a9e9-535a999fc3a4 for instance with vm_state building and task_state spawning. [ 1176.277039] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3af792c1-4116-4fb2-9f15-57153c51929f tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "90932f05-fad3-4c6d-87ae-ab059351b0be" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.571s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1176.360373] env[62692]: DEBUG nova.network.neutron [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Successfully updated port: ea6ff473-6d45-4425-a9e9-535a999fc3a4 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1176.550811] env[62692]: DEBUG nova.network.neutron [req-b6d528eb-5352-4c87-98b8-415801936619 req-9832e2a9-9ae6-4297-ba17-b54c5394c6b8 service nova] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Updated VIF entry in instance network info cache for port e2bdb2a3-7e2c-4329-9008-9bea98b0b80a. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1176.551409] env[62692]: DEBUG nova.network.neutron [req-b6d528eb-5352-4c87-98b8-415801936619 req-9832e2a9-9ae6-4297-ba17-b54c5394c6b8 service nova] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Updating instance_info_cache with network_info: [{"id": "e2bdb2a3-7e2c-4329-9008-9bea98b0b80a", "address": "fa:16:3e:3c:8b:ce", "network": {"id": "2febb660-329d-403c-8c2c-f11c0f5f547c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1806650357-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37921f67664544478cb2568def068b11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape2bdb2a3-7e", "ovs_interfaceid": "e2bdb2a3-7e2c-4329-9008-9bea98b0b80a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1176.692331] env[62692]: DEBUG oslo_vmware.api [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142010, 'name': CloneVM_Task, 'duration_secs': 2.285536} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.692659] env[62692]: INFO nova.virt.vmwareapi.vmops [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Created linked-clone VM from snapshot [ 1176.693524] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f028ee9f-12a9-49fa-b9c4-e16a8b0161ab {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.704735] env[62692]: INFO nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 0b6ea356-8353-433e-a19a-6a87dd74287a has allocations against this compute host but is not found in the database. [ 1176.704821] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 973ac4ab-5f82-4108-8076-9a370ca9d914 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1176.705187] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=62692) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1176.705310] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2432MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=62692) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1176.707569] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Uploading image c0e34614-043a-4a57-969f-c9e09d45f20e {{(pid=62692) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1176.742078] env[62692]: DEBUG oslo_vmware.rw_handles [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1176.742078] env[62692]: value = "vm-249081" [ 1176.742078] env[62692]: _type = "VirtualMachine" [ 1176.742078] env[62692]: }. {{(pid=62692) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1176.742415] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-3629cb1c-535f-41f7-83cc-427e614cd1b0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.754037] env[62692]: DEBUG oslo_vmware.rw_handles [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lease: (returnval){ [ 1176.754037] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52353090-4573-d5e4-e32d-db2312778469" [ 1176.754037] env[62692]: _type = "HttpNfcLease" [ 1176.754037] env[62692]: } obtained for exporting VM: (result){ [ 1176.754037] env[62692]: value = "vm-249081" [ 1176.754037] env[62692]: _type = "VirtualMachine" [ 1176.754037] env[62692]: }. {{(pid=62692) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1176.755409] env[62692]: DEBUG oslo_vmware.api [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the lease: (returnval){ [ 1176.755409] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52353090-4573-d5e4-e32d-db2312778469" [ 1176.755409] env[62692]: _type = "HttpNfcLease" [ 1176.755409] env[62692]: } to be ready. {{(pid=62692) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1176.765694] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1176.765694] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52353090-4573-d5e4-e32d-db2312778469" [ 1176.765694] env[62692]: _type = "HttpNfcLease" [ 1176.765694] env[62692]: } is initializing. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1176.863217] env[62692]: DEBUG oslo_concurrency.lockutils [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "refresh_cache-9d3c0909-6207-46b7-b83b-de4d7d626fd7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1176.863343] env[62692]: DEBUG oslo_concurrency.lockutils [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired lock "refresh_cache-9d3c0909-6207-46b7-b83b-de4d7d626fd7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1176.863482] env[62692]: DEBUG nova.network.neutron [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1176.931491] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dae179b6-3b93-42b1-9298-1efb32834587 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.939578] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18bc0a33-94c8-4d4b-8d5f-225ded61d44e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.969984] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-405a1323-e4b3-495d-b6ff-e4cd177b40d7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.978019] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3cedb8c-50b3-4c4d-94ee-06ec5c3d2bf9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.993739] env[62692]: DEBUG nova.compute.provider_tree [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1177.053940] env[62692]: DEBUG oslo_concurrency.lockutils [req-b6d528eb-5352-4c87-98b8-415801936619 req-9832e2a9-9ae6-4297-ba17-b54c5394c6b8 service nova] Releasing lock "refresh_cache-973ac4ab-5f82-4108-8076-9a370ca9d914" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1177.264090] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1177.264090] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52353090-4573-d5e4-e32d-db2312778469" [ 1177.264090] env[62692]: _type = "HttpNfcLease" [ 1177.264090] env[62692]: } is ready. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1177.264454] env[62692]: DEBUG oslo_vmware.rw_handles [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1177.264454] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52353090-4573-d5e4-e32d-db2312778469" [ 1177.264454] env[62692]: _type = "HttpNfcLease" [ 1177.264454] env[62692]: }. {{(pid=62692) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1177.265140] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b025993-f134-4775-b1f6-d1b8be79ac33 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.273894] env[62692]: DEBUG oslo_vmware.rw_handles [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525c9178-ef8a-369f-5618-4836b668eb8c/disk-0.vmdk from lease info. {{(pid=62692) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1177.274178] env[62692]: DEBUG oslo_vmware.rw_handles [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525c9178-ef8a-369f-5618-4836b668eb8c/disk-0.vmdk for reading. {{(pid=62692) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1177.396842] env[62692]: DEBUG nova.network.neutron [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1177.416834] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-81d70031-111b-4b67-8496-4a5d04ae1008 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.496737] env[62692]: DEBUG nova.scheduler.client.report [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1177.546994] env[62692]: DEBUG nova.network.neutron [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Updating instance_info_cache with network_info: [{"id": "ea6ff473-6d45-4425-a9e9-535a999fc3a4", "address": "fa:16:3e:10:57:4d", "network": {"id": "f70727d6-a9aa-44b0-a244-2533b3f96571", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-284105837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1037e5d436bd429391159b87a712f6e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea6ff473-6d", "ovs_interfaceid": "ea6ff473-6d45-4425-a9e9-535a999fc3a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1178.002593] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62692) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1178.003717] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.451s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1178.003717] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.352s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1178.003717] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1178.005825] env[62692]: DEBUG oslo_concurrency.lockutils [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.969s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1178.007351] env[62692]: INFO nova.compute.claims [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1178.010214] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1178.010381] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Cleaning up deleted instances {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 1178.035818] env[62692]: INFO nova.scheduler.client.report [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Deleted allocations for instance 2b21e8a1-5f5e-4f1f-bae6-512f1213eb48 [ 1178.050296] env[62692]: DEBUG oslo_concurrency.lockutils [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Releasing lock "refresh_cache-9d3c0909-6207-46b7-b83b-de4d7d626fd7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1178.050651] env[62692]: DEBUG nova.compute.manager [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Instance network_info: |[{"id": "ea6ff473-6d45-4425-a9e9-535a999fc3a4", "address": "fa:16:3e:10:57:4d", "network": {"id": "f70727d6-a9aa-44b0-a244-2533b3f96571", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-284105837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1037e5d436bd429391159b87a712f6e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea6ff473-6d", "ovs_interfaceid": "ea6ff473-6d45-4425-a9e9-535a999fc3a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1178.051282] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:10:57:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f44b2fa3-6730-4b87-8839-947eff21213f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ea6ff473-6d45-4425-a9e9-535a999fc3a4', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1178.060109] env[62692]: DEBUG oslo.service.loopingcall [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1178.060109] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1178.060109] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-50d4745e-b7d4-4d42-8c2a-302e00913e31 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.083316] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1178.083316] env[62692]: value = "task-1142013" [ 1178.083316] env[62692]: _type = "Task" [ 1178.083316] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.092589] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1142013, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.115126] env[62692]: DEBUG nova.compute.manager [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Stashing vm_state: stopped {{(pid=62692) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1178.285177] env[62692]: DEBUG nova.compute.manager [req-902f511a-dc74-4b3e-b9a4-8e64c115ff92 req-2dfc1007-7847-468a-b4f9-890a4a58f007 service nova] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Received event network-changed-ea6ff473-6d45-4425-a9e9-535a999fc3a4 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1178.285290] env[62692]: DEBUG nova.compute.manager [req-902f511a-dc74-4b3e-b9a4-8e64c115ff92 req-2dfc1007-7847-468a-b4f9-890a4a58f007 service nova] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Refreshing instance network info cache due to event network-changed-ea6ff473-6d45-4425-a9e9-535a999fc3a4. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1178.286026] env[62692]: DEBUG oslo_concurrency.lockutils [req-902f511a-dc74-4b3e-b9a4-8e64c115ff92 req-2dfc1007-7847-468a-b4f9-890a4a58f007 service nova] Acquiring lock "refresh_cache-9d3c0909-6207-46b7-b83b-de4d7d626fd7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1178.286026] env[62692]: DEBUG oslo_concurrency.lockutils [req-902f511a-dc74-4b3e-b9a4-8e64c115ff92 req-2dfc1007-7847-468a-b4f9-890a4a58f007 service nova] Acquired lock "refresh_cache-9d3c0909-6207-46b7-b83b-de4d7d626fd7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.286026] env[62692]: DEBUG nova.network.neutron [req-902f511a-dc74-4b3e-b9a4-8e64c115ff92 req-2dfc1007-7847-468a-b4f9-890a4a58f007 service nova] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Refreshing network info cache for port ea6ff473-6d45-4425-a9e9-535a999fc3a4 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1178.531510] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] There are 58 instances to clean {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1178.532968] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: c20139ae-5076-4f6b-b4f0-047332559acf] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1178.544492] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5abf2b68-0d64-448f-9121-ec7e87acff59 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Lock "2b21e8a1-5f5e-4f1f-bae6-512f1213eb48" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.892s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1178.594772] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1142013, 'name': CreateVM_Task, 'duration_secs': 0.335872} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.597878] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1178.598615] env[62692]: DEBUG oslo_concurrency.lockutils [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1178.598823] env[62692]: DEBUG oslo_concurrency.lockutils [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.599138] env[62692]: DEBUG oslo_concurrency.lockutils [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1178.599442] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a9045aa-9597-4833-82ee-60fbb5126679 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.605418] env[62692]: DEBUG oslo_vmware.api [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1178.605418] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52a3e7ce-b284-5d87-28e5-40c36788fc2a" [ 1178.605418] env[62692]: _type = "Task" [ 1178.605418] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.622391] env[62692]: DEBUG oslo_vmware.api [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52a3e7ce-b284-5d87-28e5-40c36788fc2a, 'name': SearchDatastore_Task, 'duration_secs': 0.011472} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.625322] env[62692]: DEBUG oslo_concurrency.lockutils [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1178.625656] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1178.625974] env[62692]: DEBUG oslo_concurrency.lockutils [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1178.626267] env[62692]: DEBUG oslo_concurrency.lockutils [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.626476] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1178.627352] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-30a308ea-f67e-48b9-af5c-680e376722c7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.634920] env[62692]: DEBUG oslo_concurrency.lockutils [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1178.636244] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1178.636499] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1178.639938] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0279bcd3-b479-402b-b4bc-19d2957c8d85 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.646379] env[62692]: DEBUG oslo_vmware.api [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1178.646379] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]523c4134-9512-a8da-1059-afb2605deed6" [ 1178.646379] env[62692]: _type = "Task" [ 1178.646379] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.655590] env[62692]: DEBUG oslo_vmware.api [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]523c4134-9512-a8da-1059-afb2605deed6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.758352] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0d384a3-dc1d-4bac-9c43-7f164f78c044 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.767160] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-519e882b-fef7-4706-9fc4-5e99d7f92b2c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.799544] env[62692]: DEBUG oslo_concurrency.lockutils [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquiring lock "dd8a0d53-5dc8-4b66-add7-f244e600468d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1178.799920] env[62692]: DEBUG oslo_concurrency.lockutils [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Lock "dd8a0d53-5dc8-4b66-add7-f244e600468d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1178.800186] env[62692]: DEBUG oslo_concurrency.lockutils [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquiring lock "dd8a0d53-5dc8-4b66-add7-f244e600468d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1178.800474] env[62692]: DEBUG oslo_concurrency.lockutils [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Lock "dd8a0d53-5dc8-4b66-add7-f244e600468d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1178.800878] env[62692]: DEBUG oslo_concurrency.lockutils [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Lock "dd8a0d53-5dc8-4b66-add7-f244e600468d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1178.805625] env[62692]: INFO nova.compute.manager [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Terminating instance [ 1178.808147] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7b78921-54ba-44dd-bf15-03b3d21c8cdc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.811836] env[62692]: DEBUG nova.compute.manager [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1178.812080] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1178.813418] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c925200-0e01-4734-a244-6005211fd822 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.826525] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-985ef4a2-7db3-4d98-8b49-007a5718c379 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.832536] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1178.832902] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f6f9ae24-adde-4c4e-919a-924aa95e9b41 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.846469] env[62692]: DEBUG nova.compute.provider_tree [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1178.851636] env[62692]: DEBUG oslo_vmware.api [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1178.851636] env[62692]: value = "task-1142014" [ 1178.851636] env[62692]: _type = "Task" [ 1178.851636] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.861796] env[62692]: DEBUG oslo_vmware.api [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1142014, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.039401] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: e55a1a5b-3656-48d7-a610-04610b43999a] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1179.158255] env[62692]: DEBUG oslo_vmware.api [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]523c4134-9512-a8da-1059-afb2605deed6, 'name': SearchDatastore_Task, 'duration_secs': 0.009927} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.159378] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d126965-510e-4362-aa41-4eb537995ba2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.165880] env[62692]: DEBUG oslo_vmware.api [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1179.165880] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]529cc47b-dbf3-86c7-d700-51ce445a704c" [ 1179.165880] env[62692]: _type = "Task" [ 1179.165880] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.175897] env[62692]: DEBUG oslo_vmware.api [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]529cc47b-dbf3-86c7-d700-51ce445a704c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.271032] env[62692]: DEBUG nova.network.neutron [req-902f511a-dc74-4b3e-b9a4-8e64c115ff92 req-2dfc1007-7847-468a-b4f9-890a4a58f007 service nova] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Updated VIF entry in instance network info cache for port ea6ff473-6d45-4425-a9e9-535a999fc3a4. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1179.271529] env[62692]: DEBUG nova.network.neutron [req-902f511a-dc74-4b3e-b9a4-8e64c115ff92 req-2dfc1007-7847-468a-b4f9-890a4a58f007 service nova] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Updating instance_info_cache with network_info: [{"id": "ea6ff473-6d45-4425-a9e9-535a999fc3a4", "address": "fa:16:3e:10:57:4d", "network": {"id": "f70727d6-a9aa-44b0-a244-2533b3f96571", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-284105837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1037e5d436bd429391159b87a712f6e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea6ff473-6d", "ovs_interfaceid": "ea6ff473-6d45-4425-a9e9-535a999fc3a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1179.353936] env[62692]: DEBUG nova.scheduler.client.report [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1179.367937] env[62692]: DEBUG oslo_vmware.api [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1142014, 'name': PowerOffVM_Task, 'duration_secs': 0.274478} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.368219] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1179.368398] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1179.368661] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b38a973d-bca8-48c3-9c8a-99fbecbac5d6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.447635] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1179.448031] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1179.448252] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Deleting the datastore file [datastore1] dd8a0d53-5dc8-4b66-add7-f244e600468d {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1179.448893] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d739dd7d-6162-4c92-be18-b28a7bf8885a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.456945] env[62692]: DEBUG oslo_vmware.api [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for the task: (returnval){ [ 1179.456945] env[62692]: value = "task-1142016" [ 1179.456945] env[62692]: _type = "Task" [ 1179.456945] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.465857] env[62692]: DEBUG oslo_vmware.api [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1142016, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.543859] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: adfb0194-e94b-4a87-8554-a5c262fe9a70] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1179.677492] env[62692]: DEBUG oslo_vmware.api [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]529cc47b-dbf3-86c7-d700-51ce445a704c, 'name': SearchDatastore_Task, 'duration_secs': 0.014629} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.677746] env[62692]: DEBUG oslo_concurrency.lockutils [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1179.678097] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 9d3c0909-6207-46b7-b83b-de4d7d626fd7/9d3c0909-6207-46b7-b83b-de4d7d626fd7.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1179.678482] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-04e256b8-c93e-4a1f-b953-dfa290864aae {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.685342] env[62692]: DEBUG oslo_vmware.api [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1179.685342] env[62692]: value = "task-1142017" [ 1179.685342] env[62692]: _type = "Task" [ 1179.685342] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.693598] env[62692]: DEBUG oslo_vmware.api [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142017, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.774941] env[62692]: DEBUG oslo_concurrency.lockutils [req-902f511a-dc74-4b3e-b9a4-8e64c115ff92 req-2dfc1007-7847-468a-b4f9-890a4a58f007 service nova] Releasing lock "refresh_cache-9d3c0909-6207-46b7-b83b-de4d7d626fd7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1179.860274] env[62692]: DEBUG oslo_concurrency.lockutils [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.854s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1179.860918] env[62692]: DEBUG nova.compute.manager [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1179.863736] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.091s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1179.865287] env[62692]: INFO nova.compute.claims [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1179.972184] env[62692]: DEBUG oslo_vmware.api [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Task: {'id': task-1142016, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.214575} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.972550] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1179.973075] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1179.973075] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1179.973147] env[62692]: INFO nova.compute.manager [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1179.973430] env[62692]: DEBUG oslo.service.loopingcall [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1179.973639] env[62692]: DEBUG nova.compute.manager [-] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1179.973737] env[62692]: DEBUG nova.network.neutron [-] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1180.046835] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 1904edb2-fd72-43b6-9219-8043c5abff0c] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1180.198623] env[62692]: DEBUG oslo_vmware.api [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142017, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.336062] env[62692]: DEBUG nova.compute.manager [req-407261ca-cba2-466f-9296-2476c4141e43 req-e82ae553-920b-4ef6-a7cb-b6fa9c0275f6 service nova] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Received event network-vif-deleted-f8a6a1d2-a31b-451a-854c-8efc6e7aa879 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1180.337132] env[62692]: INFO nova.compute.manager [req-407261ca-cba2-466f-9296-2476c4141e43 req-e82ae553-920b-4ef6-a7cb-b6fa9c0275f6 service nova] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Neutron deleted interface f8a6a1d2-a31b-451a-854c-8efc6e7aa879; detaching it from the instance and deleting it from the info cache [ 1180.337132] env[62692]: DEBUG nova.network.neutron [req-407261ca-cba2-466f-9296-2476c4141e43 req-e82ae553-920b-4ef6-a7cb-b6fa9c0275f6 service nova] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1180.370326] env[62692]: DEBUG nova.compute.utils [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1180.374277] env[62692]: DEBUG nova.compute.manager [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1180.374520] env[62692]: DEBUG nova.network.neutron [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1180.415453] env[62692]: DEBUG nova.policy [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '847654db6b294a828455ae22db82c50e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f94f7500bf524e3890734a58937b80c4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1180.550417] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 1d9271aa-d41d-48cd-a879-3306e7212197] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1180.701764] env[62692]: DEBUG oslo_vmware.api [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142017, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.571215} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.702831] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 9d3c0909-6207-46b7-b83b-de4d7d626fd7/9d3c0909-6207-46b7-b83b-de4d7d626fd7.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1180.703103] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1180.703410] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bc9074e5-04dc-435c-b01d-370d2a152627 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.708804] env[62692]: DEBUG nova.network.neutron [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Successfully created port: 9970a338-22fc-4fbf-9c8d-e2a446b21682 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1180.714615] env[62692]: DEBUG oslo_vmware.api [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1180.714615] env[62692]: value = "task-1142018" [ 1180.714615] env[62692]: _type = "Task" [ 1180.714615] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.723349] env[62692]: DEBUG oslo_vmware.api [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142018, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.813307] env[62692]: DEBUG nova.network.neutron [-] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1180.839354] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e1f6ce0e-cf71-4504-a8d8-7e0276f3415b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.849877] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d412957b-415e-4908-bfa2-e8d14355e002 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.881425] env[62692]: DEBUG nova.compute.manager [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1180.887274] env[62692]: DEBUG nova.compute.manager [req-407261ca-cba2-466f-9296-2476c4141e43 req-e82ae553-920b-4ef6-a7cb-b6fa9c0275f6 service nova] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Detach interface failed, port_id=f8a6a1d2-a31b-451a-854c-8efc6e7aa879, reason: Instance dd8a0d53-5dc8-4b66-add7-f244e600468d could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1181.054740] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: f92fae48-7285-4b69-97c9-2360227c18fc] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1181.102725] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04956702-c6e9-4f76-811d-b3a50fe359ae {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.112020] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f64d7f-1f69-43ee-8172-7c4088dc017e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.142979] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c06e5809-5045-4702-8129-02ea09bb7827 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.151798] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec1f189a-53e4-4886-aadb-ed5e7671feb4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.165787] env[62692]: DEBUG nova.compute.provider_tree [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1181.224880] env[62692]: DEBUG oslo_vmware.api [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142018, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068293} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.225413] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1181.226118] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed9b3b4c-ad93-4b42-8a1d-3a904ad19d65 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.249768] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] 9d3c0909-6207-46b7-b83b-de4d7d626fd7/9d3c0909-6207-46b7-b83b-de4d7d626fd7.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1181.250588] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-411f586b-2210-438f-af18-46623587e4df {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.272924] env[62692]: DEBUG oslo_vmware.api [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1181.272924] env[62692]: value = "task-1142019" [ 1181.272924] env[62692]: _type = "Task" [ 1181.272924] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.283660] env[62692]: DEBUG oslo_vmware.api [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142019, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.316025] env[62692]: INFO nova.compute.manager [-] [instance: dd8a0d53-5dc8-4b66-add7-f244e600468d] Took 1.34 seconds to deallocate network for instance. [ 1181.557964] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: f27e3a4f-d109-472b-a348-5099bdc0e7c8] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1181.668892] env[62692]: DEBUG nova.scheduler.client.report [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1181.783677] env[62692]: DEBUG oslo_vmware.api [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142019, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.823904] env[62692]: DEBUG oslo_concurrency.lockutils [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1181.894453] env[62692]: DEBUG nova.compute.manager [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1181.924428] env[62692]: DEBUG nova.virt.hardware [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1181.924739] env[62692]: DEBUG nova.virt.hardware [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1181.924912] env[62692]: DEBUG nova.virt.hardware [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1181.925126] env[62692]: DEBUG nova.virt.hardware [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1181.925285] env[62692]: DEBUG nova.virt.hardware [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1181.925440] env[62692]: DEBUG nova.virt.hardware [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1181.925663] env[62692]: DEBUG nova.virt.hardware [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1181.925849] env[62692]: DEBUG nova.virt.hardware [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1181.926028] env[62692]: DEBUG nova.virt.hardware [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1181.926201] env[62692]: DEBUG nova.virt.hardware [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1181.926380] env[62692]: DEBUG nova.virt.hardware [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1181.927302] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2282c379-90da-46e7-8763-19f4d561f330 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.936028] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a923d9-68ea-472f-acda-256e439c0ef6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.061061] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: cd2dd031-b52d-4dbe-adbe-2dc73e9e4b43] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1182.173985] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.310s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1182.174713] env[62692]: DEBUG nova.compute.manager [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1182.177611] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.352s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1182.179742] env[62692]: INFO nova.compute.claims [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1182.284799] env[62692]: DEBUG oslo_vmware.api [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142019, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.305904] env[62692]: DEBUG nova.network.neutron [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Successfully updated port: 9970a338-22fc-4fbf-9c8d-e2a446b21682 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1182.363515] env[62692]: DEBUG nova.compute.manager [req-7a99975e-6758-461f-8936-a4caffd9d152 req-adf28c51-a8a9-4e52-94c7-cd1f2b23b161 service nova] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Received event network-vif-plugged-9970a338-22fc-4fbf-9c8d-e2a446b21682 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1182.363789] env[62692]: DEBUG oslo_concurrency.lockutils [req-7a99975e-6758-461f-8936-a4caffd9d152 req-adf28c51-a8a9-4e52-94c7-cd1f2b23b161 service nova] Acquiring lock "9d754b1a-0e1b-4200-804e-9e4fa3bd06bf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1182.364091] env[62692]: DEBUG oslo_concurrency.lockutils [req-7a99975e-6758-461f-8936-a4caffd9d152 req-adf28c51-a8a9-4e52-94c7-cd1f2b23b161 service nova] Lock "9d754b1a-0e1b-4200-804e-9e4fa3bd06bf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1182.364353] env[62692]: DEBUG oslo_concurrency.lockutils [req-7a99975e-6758-461f-8936-a4caffd9d152 req-adf28c51-a8a9-4e52-94c7-cd1f2b23b161 service nova] Lock "9d754b1a-0e1b-4200-804e-9e4fa3bd06bf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1182.364569] env[62692]: DEBUG nova.compute.manager [req-7a99975e-6758-461f-8936-a4caffd9d152 req-adf28c51-a8a9-4e52-94c7-cd1f2b23b161 service nova] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] No waiting events found dispatching network-vif-plugged-9970a338-22fc-4fbf-9c8d-e2a446b21682 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1182.364802] env[62692]: WARNING nova.compute.manager [req-7a99975e-6758-461f-8936-a4caffd9d152 req-adf28c51-a8a9-4e52-94c7-cd1f2b23b161 service nova] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Received unexpected event network-vif-plugged-9970a338-22fc-4fbf-9c8d-e2a446b21682 for instance with vm_state building and task_state spawning. [ 1182.365014] env[62692]: DEBUG nova.compute.manager [req-7a99975e-6758-461f-8936-a4caffd9d152 req-adf28c51-a8a9-4e52-94c7-cd1f2b23b161 service nova] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Received event network-changed-9970a338-22fc-4fbf-9c8d-e2a446b21682 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1182.365248] env[62692]: DEBUG nova.compute.manager [req-7a99975e-6758-461f-8936-a4caffd9d152 req-adf28c51-a8a9-4e52-94c7-cd1f2b23b161 service nova] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Refreshing instance network info cache due to event network-changed-9970a338-22fc-4fbf-9c8d-e2a446b21682. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1182.365492] env[62692]: DEBUG oslo_concurrency.lockutils [req-7a99975e-6758-461f-8936-a4caffd9d152 req-adf28c51-a8a9-4e52-94c7-cd1f2b23b161 service nova] Acquiring lock "refresh_cache-9d754b1a-0e1b-4200-804e-9e4fa3bd06bf" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1182.365664] env[62692]: DEBUG oslo_concurrency.lockutils [req-7a99975e-6758-461f-8936-a4caffd9d152 req-adf28c51-a8a9-4e52-94c7-cd1f2b23b161 service nova] Acquired lock "refresh_cache-9d754b1a-0e1b-4200-804e-9e4fa3bd06bf" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1182.365837] env[62692]: DEBUG nova.network.neutron [req-7a99975e-6758-461f-8936-a4caffd9d152 req-adf28c51-a8a9-4e52-94c7-cd1f2b23b161 service nova] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Refreshing network info cache for port 9970a338-22fc-4fbf-9c8d-e2a446b21682 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1182.565080] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 94a07505-661f-4595-9024-8a9da9f2c2fc] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1182.684392] env[62692]: DEBUG nova.compute.utils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1182.687793] env[62692]: DEBUG nova.compute.manager [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1182.687918] env[62692]: DEBUG nova.network.neutron [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1182.727057] env[62692]: DEBUG nova.policy [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b7d10dd39f064a23ae4fa8a8ec777d92', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5a1ab0259e314d22b8f36ba6a9f4d693', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1182.785372] env[62692]: DEBUG oslo_vmware.api [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142019, 'name': ReconfigVM_Task, 'duration_secs': 1.302153} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.785645] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Reconfigured VM instance instance-00000072 to attach disk [datastore2] 9d3c0909-6207-46b7-b83b-de4d7d626fd7/9d3c0909-6207-46b7-b83b-de4d7d626fd7.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1182.786208] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-91ccf852-d1e9-4957-9f91-50088568fea6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.807597] env[62692]: DEBUG oslo_vmware.api [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1182.807597] env[62692]: value = "task-1142020" [ 1182.807597] env[62692]: _type = "Task" [ 1182.807597] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.811383] env[62692]: DEBUG oslo_concurrency.lockutils [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquiring lock "refresh_cache-9d754b1a-0e1b-4200-804e-9e4fa3bd06bf" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1182.818655] env[62692]: DEBUG oslo_vmware.api [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142020, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.900670] env[62692]: DEBUG nova.network.neutron [req-7a99975e-6758-461f-8936-a4caffd9d152 req-adf28c51-a8a9-4e52-94c7-cd1f2b23b161 service nova] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1182.990132] env[62692]: DEBUG nova.network.neutron [req-7a99975e-6758-461f-8936-a4caffd9d152 req-adf28c51-a8a9-4e52-94c7-cd1f2b23b161 service nova] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1183.056294] env[62692]: DEBUG nova.network.neutron [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Successfully created port: dff1785d-8de1-4070-9d59-44f56eb42854 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1183.067828] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: a5cdd714-2b8c-4cb1-9b0a-cd2fb6254970] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1183.189092] env[62692]: DEBUG nova.compute.manager [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1183.321034] env[62692]: DEBUG oslo_vmware.api [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142020, 'name': Rename_Task, 'duration_secs': 0.154234} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.321332] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1183.321616] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-db8013ae-52a5-4d2e-8b82-792bceb79a17 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.330362] env[62692]: DEBUG oslo_vmware.api [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1183.330362] env[62692]: value = "task-1142021" [ 1183.330362] env[62692]: _type = "Task" [ 1183.330362] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.342551] env[62692]: DEBUG oslo_vmware.api [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142021, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.415111] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0262bdab-909d-40f5-aa59-29684b235e99 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.425537] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-930c2e4d-d0d9-4719-ade4-6b1fb6e88025 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.455568] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f891293-14d0-41ca-a452-a3668c70ff0f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.464596] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c88a7eb8-eb22-43c0-8c2d-6721940aac63 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.480520] env[62692]: DEBUG nova.compute.provider_tree [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1183.492338] env[62692]: DEBUG oslo_concurrency.lockutils [req-7a99975e-6758-461f-8936-a4caffd9d152 req-adf28c51-a8a9-4e52-94c7-cd1f2b23b161 service nova] Releasing lock "refresh_cache-9d754b1a-0e1b-4200-804e-9e4fa3bd06bf" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1183.492741] env[62692]: DEBUG oslo_concurrency.lockutils [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquired lock "refresh_cache-9d754b1a-0e1b-4200-804e-9e4fa3bd06bf" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1183.492921] env[62692]: DEBUG nova.network.neutron [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1183.571341] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: e086d79d-29e6-41bf-a139-0b680ca0021d] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1183.841720] env[62692]: DEBUG oslo_vmware.api [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142021, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.984150] env[62692]: DEBUG nova.scheduler.client.report [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1184.024898] env[62692]: DEBUG nova.network.neutron [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1184.078274] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: ca56c2ec-95fc-44fc-856f-e395bda1703c] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1184.197978] env[62692]: DEBUG nova.network.neutron [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Updating instance_info_cache with network_info: [{"id": "9970a338-22fc-4fbf-9c8d-e2a446b21682", "address": "fa:16:3e:f5:9a:66", "network": {"id": "189f2464-2a0a-4637-b0f3-2f6ff8fdb324", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1462324860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f94f7500bf524e3890734a58937b80c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9970a338-22", "ovs_interfaceid": "9970a338-22fc-4fbf-9c8d-e2a446b21682", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1184.202082] env[62692]: DEBUG nova.compute.manager [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1184.229952] env[62692]: DEBUG nova.virt.hardware [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1184.230355] env[62692]: DEBUG nova.virt.hardware [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1184.231056] env[62692]: DEBUG nova.virt.hardware [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1184.231056] env[62692]: DEBUG nova.virt.hardware [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1184.231056] env[62692]: DEBUG nova.virt.hardware [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1184.231299] env[62692]: DEBUG nova.virt.hardware [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1184.231426] env[62692]: DEBUG nova.virt.hardware [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1184.231612] env[62692]: DEBUG nova.virt.hardware [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1184.231826] env[62692]: DEBUG nova.virt.hardware [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1184.232026] env[62692]: DEBUG nova.virt.hardware [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1184.232280] env[62692]: DEBUG nova.virt.hardware [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1184.233664] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad87b803-d387-4167-9ebd-4cf02534923d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.242904] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63c01aa8-bf24-4003-ad0d-eb8e6fd936dd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.341253] env[62692]: DEBUG oslo_vmware.api [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142021, 'name': PowerOnVM_Task, 'duration_secs': 0.543004} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.341540] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1184.341750] env[62692]: INFO nova.compute.manager [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Took 8.76 seconds to spawn the instance on the hypervisor. [ 1184.341939] env[62692]: DEBUG nova.compute.manager [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1184.342793] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a5e4672-0352-4fa7-949d-5979eeabcde2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.489769] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.312s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1184.490428] env[62692]: DEBUG nova.compute.manager [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1184.493115] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 8.611s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1184.548213] env[62692]: DEBUG nova.compute.manager [req-b7ba47c5-7df5-4e32-9879-85a7ffc55793 req-c4332356-9713-4c9c-a44d-d070f740891a service nova] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Received event network-vif-plugged-dff1785d-8de1-4070-9d59-44f56eb42854 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1184.548438] env[62692]: DEBUG oslo_concurrency.lockutils [req-b7ba47c5-7df5-4e32-9879-85a7ffc55793 req-c4332356-9713-4c9c-a44d-d070f740891a service nova] Acquiring lock "10d13b95-e6a1-4a11-9d8c-540ce82316de-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.548652] env[62692]: DEBUG oslo_concurrency.lockutils [req-b7ba47c5-7df5-4e32-9879-85a7ffc55793 req-c4332356-9713-4c9c-a44d-d070f740891a service nova] Lock "10d13b95-e6a1-4a11-9d8c-540ce82316de-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1184.548857] env[62692]: DEBUG oslo_concurrency.lockutils [req-b7ba47c5-7df5-4e32-9879-85a7ffc55793 req-c4332356-9713-4c9c-a44d-d070f740891a service nova] Lock "10d13b95-e6a1-4a11-9d8c-540ce82316de-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1184.548989] env[62692]: DEBUG nova.compute.manager [req-b7ba47c5-7df5-4e32-9879-85a7ffc55793 req-c4332356-9713-4c9c-a44d-d070f740891a service nova] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] No waiting events found dispatching network-vif-plugged-dff1785d-8de1-4070-9d59-44f56eb42854 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1184.549195] env[62692]: WARNING nova.compute.manager [req-b7ba47c5-7df5-4e32-9879-85a7ffc55793 req-c4332356-9713-4c9c-a44d-d070f740891a service nova] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Received unexpected event network-vif-plugged-dff1785d-8de1-4070-9d59-44f56eb42854 for instance with vm_state building and task_state spawning. [ 1184.581498] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: a1951dba-ec5a-4849-9e22-37fba08ec60c] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1184.640134] env[62692]: DEBUG nova.network.neutron [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Successfully updated port: dff1785d-8de1-4070-9d59-44f56eb42854 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1184.700747] env[62692]: DEBUG oslo_concurrency.lockutils [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Releasing lock "refresh_cache-9d754b1a-0e1b-4200-804e-9e4fa3bd06bf" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1184.701148] env[62692]: DEBUG nova.compute.manager [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Instance network_info: |[{"id": "9970a338-22fc-4fbf-9c8d-e2a446b21682", "address": "fa:16:3e:f5:9a:66", "network": {"id": "189f2464-2a0a-4637-b0f3-2f6ff8fdb324", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1462324860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f94f7500bf524e3890734a58937b80c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9970a338-22", "ovs_interfaceid": "9970a338-22fc-4fbf-9c8d-e2a446b21682", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1184.701638] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f5:9a:66', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '669e4919-e0ad-4e23-9f23-4c5f2be0d858', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9970a338-22fc-4fbf-9c8d-e2a446b21682', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1184.710095] env[62692]: DEBUG oslo.service.loopingcall [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1184.710398] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1184.710827] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-26395222-5798-4adb-abe2-db055451af96 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.734210] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1184.734210] env[62692]: value = "task-1142022" [ 1184.734210] env[62692]: _type = "Task" [ 1184.734210] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.743037] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1142022, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.861493] env[62692]: INFO nova.compute.manager [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Took 15.11 seconds to build instance. [ 1184.996727] env[62692]: DEBUG nova.compute.utils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1185.000665] env[62692]: INFO nova.compute.claims [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1185.004787] env[62692]: DEBUG nova.compute.manager [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1185.004787] env[62692]: DEBUG nova.network.neutron [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1185.055567] env[62692]: DEBUG nova.policy [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b7d10dd39f064a23ae4fa8a8ec777d92', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5a1ab0259e314d22b8f36ba6a9f4d693', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1185.086521] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: af260f9a-4e9b-4af1-90da-c619c4755eca] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1185.142578] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "refresh_cache-10d13b95-e6a1-4a11-9d8c-540ce82316de" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1185.142787] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquired lock "refresh_cache-10d13b95-e6a1-4a11-9d8c-540ce82316de" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.142847] env[62692]: DEBUG nova.network.neutron [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1185.245453] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1142022, 'name': CreateVM_Task, 'duration_secs': 0.398433} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.245656] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1185.246432] env[62692]: DEBUG oslo_concurrency.lockutils [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1185.246638] env[62692]: DEBUG oslo_concurrency.lockutils [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.247041] env[62692]: DEBUG oslo_concurrency.lockutils [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1185.247331] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f12f42a9-d06c-4343-b0b2-1eb9d155abf1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.256515] env[62692]: DEBUG oslo_vmware.api [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1185.256515] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52d49509-ab14-d3bf-df12-77367b8d6da2" [ 1185.256515] env[62692]: _type = "Task" [ 1185.256515] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.266441] env[62692]: DEBUG oslo_vmware.api [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52d49509-ab14-d3bf-df12-77367b8d6da2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.317701] env[62692]: DEBUG nova.network.neutron [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Successfully created port: fad50fb6-51f6-4917-beea-f67bf1334512 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1185.362940] env[62692]: DEBUG oslo_concurrency.lockutils [None req-701a5ab8-accd-494f-ac56-aede8abf2dba tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "9d3c0909-6207-46b7-b83b-de4d7d626fd7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.627s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1185.509805] env[62692]: INFO nova.compute.resource_tracker [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Updating resource usage from migration 0b6ea356-8353-433e-a19a-6a87dd74287a [ 1185.512609] env[62692]: DEBUG nova.compute.manager [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1185.591237] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 9e3b330d-c4ca-4feb-9ebf-5b2334e3d5da] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1185.708314] env[62692]: DEBUG nova.network.neutron [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1185.770976] env[62692]: DEBUG oslo_vmware.api [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52d49509-ab14-d3bf-df12-77367b8d6da2, 'name': SearchDatastore_Task, 'duration_secs': 0.014935} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.770976] env[62692]: DEBUG oslo_concurrency.lockutils [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1185.770976] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1185.771186] env[62692]: DEBUG oslo_concurrency.lockutils [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1185.771335] env[62692]: DEBUG oslo_concurrency.lockutils [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.771519] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1185.771807] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ab478a45-70f4-4e2f-9708-c1d5aa418693 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.783375] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1185.783584] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1185.784371] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ae52024-58bb-4b3d-9b16-75e10c38ea15 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.794885] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12993170-673d-4e07-b96b-17f5e64dfd6e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.807235] env[62692]: DEBUG oslo_vmware.api [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1185.807235] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52a4de50-bfcd-6d2c-96e3-d1d3db3d3fad" [ 1185.807235] env[62692]: _type = "Task" [ 1185.807235] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.815643] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6156bba8-050e-4f74-b9e8-b1ead4181b0e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.823610] env[62692]: DEBUG oslo_vmware.api [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52a4de50-bfcd-6d2c-96e3-d1d3db3d3fad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.876181] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0067f25-c94f-4b27-b51c-e6492f955a7b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.893475] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc8e5d7b-ffb5-4c02-9bc3-c2e62e57d551 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.914738] env[62692]: DEBUG nova.compute.provider_tree [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1186.094872] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: e8f645da-3b4a-4f21-9a17-aebdc8d112df] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1186.279901] env[62692]: DEBUG nova.network.neutron [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Updating instance_info_cache with network_info: [{"id": "dff1785d-8de1-4070-9d59-44f56eb42854", "address": "fa:16:3e:44:f4:32", "network": {"id": "8d7f8236-a3f7-4ce3-a9b4-c90747b73b30", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-289766533-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a1ab0259e314d22b8f36ba6a9f4d693", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdff1785d-8d", "ovs_interfaceid": "dff1785d-8de1-4070-9d59-44f56eb42854", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1186.318217] env[62692]: DEBUG oslo_vmware.api [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52a4de50-bfcd-6d2c-96e3-d1d3db3d3fad, 'name': SearchDatastore_Task, 'duration_secs': 0.016461} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.319101] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39aa48d4-f160-4785-acfc-4df8b9707c4f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.325741] env[62692]: DEBUG oslo_vmware.api [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1186.325741] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5290620a-9627-775a-d402-cedd74e79051" [ 1186.325741] env[62692]: _type = "Task" [ 1186.325741] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.335913] env[62692]: DEBUG oslo_vmware.api [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5290620a-9627-775a-d402-cedd74e79051, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.411894] env[62692]: DEBUG nova.compute.manager [req-e7683ce9-c8fe-47ec-98e1-5161a66254f5 req-5aa285b3-c5b9-4ab3-b475-3844d0a537eb service nova] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Received event network-changed-ea6ff473-6d45-4425-a9e9-535a999fc3a4 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1186.412137] env[62692]: DEBUG nova.compute.manager [req-e7683ce9-c8fe-47ec-98e1-5161a66254f5 req-5aa285b3-c5b9-4ab3-b475-3844d0a537eb service nova] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Refreshing instance network info cache due to event network-changed-ea6ff473-6d45-4425-a9e9-535a999fc3a4. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1186.412415] env[62692]: DEBUG oslo_concurrency.lockutils [req-e7683ce9-c8fe-47ec-98e1-5161a66254f5 req-5aa285b3-c5b9-4ab3-b475-3844d0a537eb service nova] Acquiring lock "refresh_cache-9d3c0909-6207-46b7-b83b-de4d7d626fd7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1186.412573] env[62692]: DEBUG oslo_concurrency.lockutils [req-e7683ce9-c8fe-47ec-98e1-5161a66254f5 req-5aa285b3-c5b9-4ab3-b475-3844d0a537eb service nova] Acquired lock "refresh_cache-9d3c0909-6207-46b7-b83b-de4d7d626fd7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1186.412761] env[62692]: DEBUG nova.network.neutron [req-e7683ce9-c8fe-47ec-98e1-5161a66254f5 req-5aa285b3-c5b9-4ab3-b475-3844d0a537eb service nova] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Refreshing network info cache for port ea6ff473-6d45-4425-a9e9-535a999fc3a4 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1186.417438] env[62692]: DEBUG nova.scheduler.client.report [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1186.522191] env[62692]: DEBUG nova.compute.manager [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1186.552048] env[62692]: DEBUG nova.virt.hardware [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1186.552048] env[62692]: DEBUG nova.virt.hardware [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1186.552048] env[62692]: DEBUG nova.virt.hardware [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1186.552048] env[62692]: DEBUG nova.virt.hardware [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1186.552048] env[62692]: DEBUG nova.virt.hardware [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1186.552048] env[62692]: DEBUG nova.virt.hardware [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1186.552048] env[62692]: DEBUG nova.virt.hardware [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1186.552048] env[62692]: DEBUG nova.virt.hardware [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1186.552545] env[62692]: DEBUG nova.virt.hardware [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1186.552545] env[62692]: DEBUG nova.virt.hardware [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1186.552617] env[62692]: DEBUG nova.virt.hardware [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1186.553650] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a62e632e-66e3-4394-9288-aef29b52b821 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.564493] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a31bacfe-73ae-4560-ae90-729240796f93 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.581825] env[62692]: DEBUG nova.compute.manager [req-5df3a7de-8d41-40f8-a138-c81f26eaee9c req-4fd3e50a-9537-408a-bdb7-d57f61af103b service nova] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Received event network-changed-dff1785d-8de1-4070-9d59-44f56eb42854 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1186.582088] env[62692]: DEBUG nova.compute.manager [req-5df3a7de-8d41-40f8-a138-c81f26eaee9c req-4fd3e50a-9537-408a-bdb7-d57f61af103b service nova] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Refreshing instance network info cache due to event network-changed-dff1785d-8de1-4070-9d59-44f56eb42854. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1186.582322] env[62692]: DEBUG oslo_concurrency.lockutils [req-5df3a7de-8d41-40f8-a138-c81f26eaee9c req-4fd3e50a-9537-408a-bdb7-d57f61af103b service nova] Acquiring lock "refresh_cache-10d13b95-e6a1-4a11-9d8c-540ce82316de" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1186.600614] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 1b753882-9118-479c-aa3e-cbb8bdf2e086] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1186.615017] env[62692]: DEBUG oslo_vmware.rw_handles [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525c9178-ef8a-369f-5618-4836b668eb8c/disk-0.vmdk. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1186.615017] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46f8bff4-a0f4-4075-b821-8d7613ffaf47 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.622810] env[62692]: DEBUG oslo_vmware.rw_handles [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525c9178-ef8a-369f-5618-4836b668eb8c/disk-0.vmdk is in state: ready. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1186.623260] env[62692]: ERROR oslo_vmware.rw_handles [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525c9178-ef8a-369f-5618-4836b668eb8c/disk-0.vmdk due to incomplete transfer. [ 1186.623624] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-ce51a1de-22a3-4e71-86ef-64f666b6ec7b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.633150] env[62692]: DEBUG oslo_vmware.rw_handles [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525c9178-ef8a-369f-5618-4836b668eb8c/disk-0.vmdk. {{(pid=62692) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1186.633754] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Uploaded image c0e34614-043a-4a57-969f-c9e09d45f20e to the Glance image server {{(pid=62692) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1186.639991] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Destroying the VM {{(pid=62692) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1186.639991] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-11350267-c170-41c9-bfec-d07d76715752 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.651278] env[62692]: DEBUG oslo_vmware.api [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1186.651278] env[62692]: value = "task-1142023" [ 1186.651278] env[62692]: _type = "Task" [ 1186.651278] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.666618] env[62692]: DEBUG oslo_vmware.api [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142023, 'name': Destroy_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.782821] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Releasing lock "refresh_cache-10d13b95-e6a1-4a11-9d8c-540ce82316de" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1186.786439] env[62692]: DEBUG nova.compute.manager [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Instance network_info: |[{"id": "dff1785d-8de1-4070-9d59-44f56eb42854", "address": "fa:16:3e:44:f4:32", "network": {"id": "8d7f8236-a3f7-4ce3-a9b4-c90747b73b30", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-289766533-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a1ab0259e314d22b8f36ba6a9f4d693", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdff1785d-8d", "ovs_interfaceid": "dff1785d-8de1-4070-9d59-44f56eb42854", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1186.786439] env[62692]: DEBUG oslo_concurrency.lockutils [req-5df3a7de-8d41-40f8-a138-c81f26eaee9c req-4fd3e50a-9537-408a-bdb7-d57f61af103b service nova] Acquired lock "refresh_cache-10d13b95-e6a1-4a11-9d8c-540ce82316de" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1186.786439] env[62692]: DEBUG nova.network.neutron [req-5df3a7de-8d41-40f8-a138-c81f26eaee9c req-4fd3e50a-9537-408a-bdb7-d57f61af103b service nova] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Refreshing network info cache for port dff1785d-8de1-4070-9d59-44f56eb42854 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1186.786895] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:44:f4:32', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c1520c99-af74-4d61-a8ae-56aef56ef4f0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dff1785d-8de1-4070-9d59-44f56eb42854', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1186.798072] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Creating folder: Project (5a1ab0259e314d22b8f36ba6a9f4d693). Parent ref: group-v248868. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1186.801849] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-176ded73-22b4-472c-ab6f-e9d2f8bfee98 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.816405] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Created folder: Project (5a1ab0259e314d22b8f36ba6a9f4d693) in parent group-v248868. [ 1186.816665] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Creating folder: Instances. Parent ref: group-v249084. {{(pid=62692) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1186.816975] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-67ab2cd5-76db-4283-a8a6-8eea21c9b4e4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.834429] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Created folder: Instances in parent group-v249084. [ 1186.835191] env[62692]: DEBUG oslo.service.loopingcall [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1186.835741] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1186.835995] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d26f799b-05bd-49f2-bfbc-223e77815159 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.862273] env[62692]: DEBUG oslo_vmware.api [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5290620a-9627-775a-d402-cedd74e79051, 'name': SearchDatastore_Task, 'duration_secs': 0.029226} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.863477] env[62692]: DEBUG oslo_concurrency.lockutils [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1186.863921] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf/9d754b1a-0e1b-4200-804e-9e4fa3bd06bf.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1186.864233] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ed573f59-6373-495f-91aa-84151194a635 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.868825] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1186.868825] env[62692]: value = "task-1142026" [ 1186.868825] env[62692]: _type = "Task" [ 1186.868825] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.875314] env[62692]: DEBUG oslo_vmware.api [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1186.875314] env[62692]: value = "task-1142027" [ 1186.875314] env[62692]: _type = "Task" [ 1186.875314] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.883569] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1142026, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.889976] env[62692]: DEBUG oslo_vmware.api [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142027, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.923101] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.430s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1186.923383] env[62692]: INFO nova.compute.manager [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Migrating [ 1186.932968] env[62692]: DEBUG oslo_concurrency.lockutils [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 8.298s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1187.103386] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: f523c482-8365-47b8-9398-a56a98fcffe8] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1187.130909] env[62692]: DEBUG nova.network.neutron [req-5df3a7de-8d41-40f8-a138-c81f26eaee9c req-4fd3e50a-9537-408a-bdb7-d57f61af103b service nova] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Updated VIF entry in instance network info cache for port dff1785d-8de1-4070-9d59-44f56eb42854. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1187.131525] env[62692]: DEBUG nova.network.neutron [req-5df3a7de-8d41-40f8-a138-c81f26eaee9c req-4fd3e50a-9537-408a-bdb7-d57f61af103b service nova] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Updating instance_info_cache with network_info: [{"id": "dff1785d-8de1-4070-9d59-44f56eb42854", "address": "fa:16:3e:44:f4:32", "network": {"id": "8d7f8236-a3f7-4ce3-a9b4-c90747b73b30", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-289766533-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a1ab0259e314d22b8f36ba6a9f4d693", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdff1785d-8d", "ovs_interfaceid": "dff1785d-8de1-4070-9d59-44f56eb42854", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1187.163808] env[62692]: DEBUG oslo_vmware.api [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142023, 'name': Destroy_Task} progress is 33%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.210410] env[62692]: DEBUG nova.network.neutron [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Successfully updated port: fad50fb6-51f6-4917-beea-f67bf1334512 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1187.381997] env[62692]: DEBUG nova.network.neutron [req-e7683ce9-c8fe-47ec-98e1-5161a66254f5 req-5aa285b3-c5b9-4ab3-b475-3844d0a537eb service nova] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Updated VIF entry in instance network info cache for port ea6ff473-6d45-4425-a9e9-535a999fc3a4. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1187.382575] env[62692]: DEBUG nova.network.neutron [req-e7683ce9-c8fe-47ec-98e1-5161a66254f5 req-5aa285b3-c5b9-4ab3-b475-3844d0a537eb service nova] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Updating instance_info_cache with network_info: [{"id": "ea6ff473-6d45-4425-a9e9-535a999fc3a4", "address": "fa:16:3e:10:57:4d", "network": {"id": "f70727d6-a9aa-44b0-a244-2533b3f96571", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-284105837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1037e5d436bd429391159b87a712f6e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea6ff473-6d", "ovs_interfaceid": "ea6ff473-6d45-4425-a9e9-535a999fc3a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1187.387864] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1142026, 'name': CreateVM_Task, 'duration_secs': 0.508845} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.388736] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1187.389570] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1187.389680] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1187.390019] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1187.390337] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72425152-4b1c-41a7-9c66-5240f03b58d4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.396852] env[62692]: DEBUG oslo_vmware.api [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142027, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.398422] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1187.398422] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5207e262-265f-e4e7-f5c8-add5610b8843" [ 1187.398422] env[62692]: _type = "Task" [ 1187.398422] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.409760] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5207e262-265f-e4e7-f5c8-add5610b8843, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.447253] env[62692]: INFO nova.compute.claims [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1187.451532] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "refresh_cache-973ac4ab-5f82-4108-8076-9a370ca9d914" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1187.451681] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquired lock "refresh_cache-973ac4ab-5f82-4108-8076-9a370ca9d914" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1187.451833] env[62692]: DEBUG nova.network.neutron [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1187.608029] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: f6085887-a5d3-4096-be38-d843bb54304a] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1187.634741] env[62692]: DEBUG oslo_concurrency.lockutils [req-5df3a7de-8d41-40f8-a138-c81f26eaee9c req-4fd3e50a-9537-408a-bdb7-d57f61af103b service nova] Releasing lock "refresh_cache-10d13b95-e6a1-4a11-9d8c-540ce82316de" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1187.661840] env[62692]: DEBUG oslo_vmware.api [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142023, 'name': Destroy_Task, 'duration_secs': 0.607435} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.662102] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Destroyed the VM [ 1187.662383] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Deleting Snapshot of the VM instance {{(pid=62692) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1187.662633] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5d6bc9c2-d83b-47db-ba3f-10f5dca37010 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.670669] env[62692]: DEBUG oslo_vmware.api [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1187.670669] env[62692]: value = "task-1142028" [ 1187.670669] env[62692]: _type = "Task" [ 1187.670669] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.679804] env[62692]: DEBUG oslo_vmware.api [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142028, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.715202] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "refresh_cache-c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1187.715333] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquired lock "refresh_cache-c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1187.715553] env[62692]: DEBUG nova.network.neutron [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1187.887275] env[62692]: DEBUG oslo_vmware.api [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142027, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.659076} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.887634] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf/9d754b1a-0e1b-4200-804e-9e4fa3bd06bf.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1187.887905] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1187.888210] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a82a4f3d-6ac1-4c8d-bdfb-415d2fce3f0a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.890438] env[62692]: DEBUG oslo_concurrency.lockutils [req-e7683ce9-c8fe-47ec-98e1-5161a66254f5 req-5aa285b3-c5b9-4ab3-b475-3844d0a537eb service nova] Releasing lock "refresh_cache-9d3c0909-6207-46b7-b83b-de4d7d626fd7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1187.896301] env[62692]: DEBUG oslo_vmware.api [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1187.896301] env[62692]: value = "task-1142029" [ 1187.896301] env[62692]: _type = "Task" [ 1187.896301] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.908877] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5207e262-265f-e4e7-f5c8-add5610b8843, 'name': SearchDatastore_Task, 'duration_secs': 0.055722} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.911694] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1187.911943] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1187.912191] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1187.912383] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1187.912557] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1187.912824] env[62692]: DEBUG oslo_vmware.api [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142029, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.913050] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-07a65f13-6e1d-475f-bb0f-4d6c51c6a54b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.921562] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1187.921743] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1187.922532] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3ade20a-b627-4ee9-a773-c8dfea445f1c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.929274] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1187.929274] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52c3da6d-7a28-a053-d5ef-85134b501185" [ 1187.929274] env[62692]: _type = "Task" [ 1187.929274] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.936684] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52c3da6d-7a28-a053-d5ef-85134b501185, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.955416] env[62692]: INFO nova.compute.resource_tracker [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Updating resource usage from migration 1ab606dc-82f8-4a56-88d8-27a18eba60fc [ 1188.110660] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 7183ec4a-992d-48f1-8dda-7f499c2f4e1d] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1188.167436] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-136c6ac8-d9d4-49a7-9c0b-f59dffa4550c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.179629] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1ff5cf9-b640-4afa-9581-17fcf8f15acf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.188380] env[62692]: DEBUG oslo_vmware.api [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142028, 'name': RemoveSnapshot_Task, 'duration_secs': 0.471452} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.189183] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Deleted Snapshot of the VM instance {{(pid=62692) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1188.189656] env[62692]: DEBUG nova.compute.manager [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1188.214221] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ec77391-6f5a-4857-9fb6-be3ca59880b1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.219735] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-972b9ab5-fae0-4133-ab58-77b387aeb80f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.231964] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb9c8da7-3945-4984-8cdd-d925a3929500 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.246298] env[62692]: DEBUG nova.compute.provider_tree [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1188.261065] env[62692]: DEBUG nova.network.neutron [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Updating instance_info_cache with network_info: [{"id": "e2bdb2a3-7e2c-4329-9008-9bea98b0b80a", "address": "fa:16:3e:3c:8b:ce", "network": {"id": "2febb660-329d-403c-8c2c-f11c0f5f547c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1806650357-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37921f67664544478cb2568def068b11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape2bdb2a3-7e", "ovs_interfaceid": "e2bdb2a3-7e2c-4329-9008-9bea98b0b80a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1188.273284] env[62692]: DEBUG nova.network.neutron [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1188.407133] env[62692]: DEBUG oslo_vmware.api [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142029, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08216} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.407564] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1188.408546] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e557074-65d9-417a-b1fc-8958a1ea9559 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.438962] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Reconfiguring VM instance instance-00000073 to attach disk [datastore1] 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf/9d754b1a-0e1b-4200-804e-9e4fa3bd06bf.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1188.439352] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-572403b3-5942-4b6d-928b-b0fed95cdc91 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.465873] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52c3da6d-7a28-a053-d5ef-85134b501185, 'name': SearchDatastore_Task, 'duration_secs': 0.018272} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.467938] env[62692]: DEBUG oslo_vmware.api [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1188.467938] env[62692]: value = "task-1142030" [ 1188.467938] env[62692]: _type = "Task" [ 1188.467938] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.468189] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18d7b951-b053-4523-a440-dfcee072b7fa {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.476029] env[62692]: DEBUG nova.network.neutron [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Updating instance_info_cache with network_info: [{"id": "fad50fb6-51f6-4917-beea-f67bf1334512", "address": "fa:16:3e:fc:3a:0b", "network": {"id": "8d7f8236-a3f7-4ce3-a9b4-c90747b73b30", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-289766533-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a1ab0259e314d22b8f36ba6a9f4d693", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfad50fb6-51", "ovs_interfaceid": "fad50fb6-51f6-4917-beea-f67bf1334512", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1188.481944] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1188.481944] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52074aad-c440-ba0c-7aa1-d4c162e55e26" [ 1188.481944] env[62692]: _type = "Task" [ 1188.481944] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.482307] env[62692]: DEBUG oslo_vmware.api [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142030, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.493221] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52074aad-c440-ba0c-7aa1-d4c162e55e26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.607307] env[62692]: DEBUG nova.compute.manager [req-81a9e838-4985-4be2-94b5-0892a284b968 req-1c5394a3-bf1a-4ace-9e7d-c8cbf8577045 service nova] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Received event network-vif-plugged-fad50fb6-51f6-4917-beea-f67bf1334512 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1188.607532] env[62692]: DEBUG oslo_concurrency.lockutils [req-81a9e838-4985-4be2-94b5-0892a284b968 req-1c5394a3-bf1a-4ace-9e7d-c8cbf8577045 service nova] Acquiring lock "c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.607744] env[62692]: DEBUG oslo_concurrency.lockutils [req-81a9e838-4985-4be2-94b5-0892a284b968 req-1c5394a3-bf1a-4ace-9e7d-c8cbf8577045 service nova] Lock "c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1188.607916] env[62692]: DEBUG oslo_concurrency.lockutils [req-81a9e838-4985-4be2-94b5-0892a284b968 req-1c5394a3-bf1a-4ace-9e7d-c8cbf8577045 service nova] Lock "c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1188.608126] env[62692]: DEBUG nova.compute.manager [req-81a9e838-4985-4be2-94b5-0892a284b968 req-1c5394a3-bf1a-4ace-9e7d-c8cbf8577045 service nova] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] No waiting events found dispatching network-vif-plugged-fad50fb6-51f6-4917-beea-f67bf1334512 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1188.608360] env[62692]: WARNING nova.compute.manager [req-81a9e838-4985-4be2-94b5-0892a284b968 req-1c5394a3-bf1a-4ace-9e7d-c8cbf8577045 service nova] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Received unexpected event network-vif-plugged-fad50fb6-51f6-4917-beea-f67bf1334512 for instance with vm_state building and task_state spawning. [ 1188.608532] env[62692]: DEBUG nova.compute.manager [req-81a9e838-4985-4be2-94b5-0892a284b968 req-1c5394a3-bf1a-4ace-9e7d-c8cbf8577045 service nova] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Received event network-changed-fad50fb6-51f6-4917-beea-f67bf1334512 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1188.608693] env[62692]: DEBUG nova.compute.manager [req-81a9e838-4985-4be2-94b5-0892a284b968 req-1c5394a3-bf1a-4ace-9e7d-c8cbf8577045 service nova] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Refreshing instance network info cache due to event network-changed-fad50fb6-51f6-4917-beea-f67bf1334512. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1188.608876] env[62692]: DEBUG oslo_concurrency.lockutils [req-81a9e838-4985-4be2-94b5-0892a284b968 req-1c5394a3-bf1a-4ace-9e7d-c8cbf8577045 service nova] Acquiring lock "refresh_cache-c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1188.614088] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 16790505-cdf6-4937-8839-fa685a5f413e] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1188.735061] env[62692]: INFO nova.compute.manager [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Shelve offloading [ 1188.736979] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1188.737288] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-93b49622-c70c-4e53-b61a-a00d700e6b0a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.746400] env[62692]: DEBUG oslo_vmware.api [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1188.746400] env[62692]: value = "task-1142031" [ 1188.746400] env[62692]: _type = "Task" [ 1188.746400] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.750479] env[62692]: DEBUG nova.scheduler.client.report [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1188.760412] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] VM already powered off {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1188.760639] env[62692]: DEBUG nova.compute.manager [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1188.761500] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce32c81b-118a-452c-8a29-f4e2f479e306 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.765281] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Releasing lock "refresh_cache-973ac4ab-5f82-4108-8076-9a370ca9d914" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1188.770472] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "refresh_cache-c2f2a485-363e-41af-9b3f-37092947ca08" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1188.770680] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquired lock "refresh_cache-c2f2a485-363e-41af-9b3f-37092947ca08" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1188.770894] env[62692]: DEBUG nova.network.neutron [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1188.981029] env[62692]: DEBUG oslo_vmware.api [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142030, 'name': ReconfigVM_Task, 'duration_secs': 0.291477} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.981253] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Reconfigured VM instance instance-00000073 to attach disk [datastore1] 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf/9d754b1a-0e1b-4200-804e-9e4fa3bd06bf.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1188.981923] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3bdf85ed-e3ff-473e-bc03-0d1ca1a9d97b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.983706] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Releasing lock "refresh_cache-c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1188.983992] env[62692]: DEBUG nova.compute.manager [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Instance network_info: |[{"id": "fad50fb6-51f6-4917-beea-f67bf1334512", "address": "fa:16:3e:fc:3a:0b", "network": {"id": "8d7f8236-a3f7-4ce3-a9b4-c90747b73b30", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-289766533-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a1ab0259e314d22b8f36ba6a9f4d693", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfad50fb6-51", "ovs_interfaceid": "fad50fb6-51f6-4917-beea-f67bf1334512", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1188.984344] env[62692]: DEBUG oslo_concurrency.lockutils [req-81a9e838-4985-4be2-94b5-0892a284b968 req-1c5394a3-bf1a-4ace-9e7d-c8cbf8577045 service nova] Acquired lock "refresh_cache-c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1188.984527] env[62692]: DEBUG nova.network.neutron [req-81a9e838-4985-4be2-94b5-0892a284b968 req-1c5394a3-bf1a-4ace-9e7d-c8cbf8577045 service nova] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Refreshing network info cache for port fad50fb6-51f6-4917-beea-f67bf1334512 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1188.985746] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:3a:0b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c1520c99-af74-4d61-a8ae-56aef56ef4f0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fad50fb6-51f6-4917-beea-f67bf1334512', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1188.992866] env[62692]: DEBUG oslo.service.loopingcall [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1188.996887] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1188.997119] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f60019e4-88e4-41c4-a4e7-50a840970522 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.013121] env[62692]: DEBUG oslo_vmware.api [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1189.013121] env[62692]: value = "task-1142032" [ 1189.013121] env[62692]: _type = "Task" [ 1189.013121] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.017321] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52074aad-c440-ba0c-7aa1-d4c162e55e26, 'name': SearchDatastore_Task, 'duration_secs': 0.022655} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.020190] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1189.020479] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 10d13b95-e6a1-4a11-9d8c-540ce82316de/10d13b95-e6a1-4a11-9d8c-540ce82316de.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1189.020732] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8c5b9adb-3e11-4235-bd3c-b7c929ec6f06 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.025674] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1189.025674] env[62692]: value = "task-1142033" [ 1189.025674] env[62692]: _type = "Task" [ 1189.025674] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.031971] env[62692]: DEBUG oslo_vmware.api [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142032, 'name': Rename_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.033595] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1189.033595] env[62692]: value = "task-1142034" [ 1189.033595] env[62692]: _type = "Task" [ 1189.033595] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.040028] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1142033, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.045226] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142034, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.117409] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: f7664593-1b7e-495f-8d0e-9137d7f180cc] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1189.256067] env[62692]: DEBUG oslo_concurrency.lockutils [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.323s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1189.256408] env[62692]: INFO nova.compute.manager [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Migrating [ 1189.269068] env[62692]: DEBUG oslo_concurrency.lockutils [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.445s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1189.269393] env[62692]: DEBUG nova.objects.instance [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Lazy-loading 'resources' on Instance uuid dd8a0d53-5dc8-4b66-add7-f244e600468d {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1189.527427] env[62692]: DEBUG oslo_vmware.api [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142032, 'name': Rename_Task, 'duration_secs': 0.156804} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.530543] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1189.530818] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9b5b13ab-62ac-4ee5-b73f-cff1ecdac135 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.540246] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1142033, 'name': CreateVM_Task, 'duration_secs': 0.340355} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.542948] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1189.543316] env[62692]: DEBUG oslo_vmware.api [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1189.543316] env[62692]: value = "task-1142035" [ 1189.543316] env[62692]: _type = "Task" [ 1189.543316] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.543950] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1189.544131] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1189.544446] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1189.546930] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-663779bd-1e9d-4f85-9f02-5af559ee8e6d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.555803] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142034, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.5079} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.561180] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] 10d13b95-e6a1-4a11-9d8c-540ce82316de/10d13b95-e6a1-4a11-9d8c-540ce82316de.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1189.561571] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1189.561907] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1189.561907] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5283bff8-2fce-8711-3bfa-8e2b8ba37504" [ 1189.561907] env[62692]: _type = "Task" [ 1189.561907] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.562121] env[62692]: DEBUG oslo_vmware.api [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142035, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.562377] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3137879b-5285-4e74-b250-0af111415779 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.577013] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5283bff8-2fce-8711-3bfa-8e2b8ba37504, 'name': SearchDatastore_Task, 'duration_secs': 0.018389} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.578403] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1189.578660] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1189.578895] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1189.579058] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1189.579243] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1189.579562] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1189.579562] env[62692]: value = "task-1142036" [ 1189.579562] env[62692]: _type = "Task" [ 1189.579562] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.579750] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8d60816b-664a-4779-a316-5aeba1bf257f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.591258] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142036, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.592754] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1189.592936] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1189.593662] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34a341f9-e65f-4d26-9ad0-9f810e573e8b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.600097] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1189.600097] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52def0ad-3175-9cb4-1ed1-cc33cb5e6575" [ 1189.600097] env[62692]: _type = "Task" [ 1189.600097] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.608338] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52def0ad-3175-9cb4-1ed1-cc33cb5e6575, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.621085] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: f3e8be1c-d952-49b7-86a8-d0a6c56ee096] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1189.701728] env[62692]: DEBUG nova.network.neutron [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Updating instance_info_cache with network_info: [{"id": "fa8d79b4-2bf8-4b6b-a5c2-6893f618f420", "address": "fa:16:3e:a0:b7:ee", "network": {"id": "22795f7f-18d7-4b10-8051-bcff4e0af3ac", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-445901728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.250", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96b921004f8a4bf68640c3810fe7b819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa8d79b4-2b", "ovs_interfaceid": "fa8d79b4-2bf8-4b6b-a5c2-6893f618f420", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1189.792130] env[62692]: DEBUG oslo_concurrency.lockutils [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "refresh_cache-90932f05-fad3-4c6d-87ae-ab059351b0be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1189.792488] env[62692]: DEBUG oslo_concurrency.lockutils [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquired lock "refresh_cache-90932f05-fad3-4c6d-87ae-ab059351b0be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1189.792749] env[62692]: DEBUG nova.network.neutron [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1189.823352] env[62692]: DEBUG nova.network.neutron [req-81a9e838-4985-4be2-94b5-0892a284b968 req-1c5394a3-bf1a-4ace-9e7d-c8cbf8577045 service nova] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Updated VIF entry in instance network info cache for port fad50fb6-51f6-4917-beea-f67bf1334512. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1189.823812] env[62692]: DEBUG nova.network.neutron [req-81a9e838-4985-4be2-94b5-0892a284b968 req-1c5394a3-bf1a-4ace-9e7d-c8cbf8577045 service nova] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Updating instance_info_cache with network_info: [{"id": "fad50fb6-51f6-4917-beea-f67bf1334512", "address": "fa:16:3e:fc:3a:0b", "network": {"id": "8d7f8236-a3f7-4ce3-a9b4-c90747b73b30", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-289766533-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a1ab0259e314d22b8f36ba6a9f4d693", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfad50fb6-51", "ovs_interfaceid": "fad50fb6-51f6-4917-beea-f67bf1334512", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1189.983736] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98db028a-7e81-4ddd-801c-117aa815f51d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.992673] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d542a0e1-bc5d-448f-9184-b1a50e8d538c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.024033] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-908630dd-9e7f-49e4-96cc-6805501a7fe2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.031995] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42b1b524-e392-44df-9e88-17a55750f196 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.045754] env[62692]: DEBUG nova.compute.provider_tree [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1190.056494] env[62692]: DEBUG oslo_vmware.api [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142035, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.091700] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142036, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079477} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.091964] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1190.092796] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e27756e-0265-4848-b9ac-b2ffafcee4d0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.116040] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Reconfiguring VM instance instance-00000074 to attach disk [datastore1] 10d13b95-e6a1-4a11-9d8c-540ce82316de/10d13b95-e6a1-4a11-9d8c-540ce82316de.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1190.118865] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6563389d-a8f9-4fca-954d-3912339cb6ab {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.132843] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 668fb0ea-7eed-4198-943c-8f916ec11368] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1190.142219] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52def0ad-3175-9cb4-1ed1-cc33cb5e6575, 'name': SearchDatastore_Task, 'duration_secs': 0.02171} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.143962] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1190.143962] env[62692]: value = "task-1142037" [ 1190.143962] env[62692]: _type = "Task" [ 1190.143962] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.144181] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eee9e6e0-1682-4d1c-8aac-1d93a68aec3a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.152757] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1190.152757] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]525037fd-9020-6022-c90e-630f741a4e24" [ 1190.152757] env[62692]: _type = "Task" [ 1190.152757] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.155626] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142037, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.163270] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]525037fd-9020-6022-c90e-630f741a4e24, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.204961] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Releasing lock "refresh_cache-c2f2a485-363e-41af-9b3f-37092947ca08" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1190.303585] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5221fbcf-d031-46ba-b62d-f7d095c1cee0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.323762] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Updating instance '973ac4ab-5f82-4108-8076-9a370ca9d914' progress to 0 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1190.329021] env[62692]: DEBUG oslo_concurrency.lockutils [req-81a9e838-4985-4be2-94b5-0892a284b968 req-1c5394a3-bf1a-4ace-9e7d-c8cbf8577045 service nova] Releasing lock "refresh_cache-c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1190.549306] env[62692]: DEBUG nova.scheduler.client.report [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1190.558567] env[62692]: DEBUG oslo_vmware.api [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142035, 'name': PowerOnVM_Task, 'duration_secs': 0.555532} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.558880] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1190.559104] env[62692]: INFO nova.compute.manager [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Took 8.66 seconds to spawn the instance on the hypervisor. [ 1190.559294] env[62692]: DEBUG nova.compute.manager [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1190.560224] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11531eff-1db4-4b88-a259-463f9fddc623 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.636928] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 084ce29c-302a-45cc-b36f-0143eccee298] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1190.639665] env[62692]: DEBUG nova.compute.manager [req-4b1d3b55-ae3f-40f9-ad43-a8c1afb0e6bc req-67897676-6cc0-42e4-96b9-6f0742f4d689 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Received event network-vif-unplugged-fa8d79b4-2bf8-4b6b-a5c2-6893f618f420 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1190.639880] env[62692]: DEBUG oslo_concurrency.lockutils [req-4b1d3b55-ae3f-40f9-ad43-a8c1afb0e6bc req-67897676-6cc0-42e4-96b9-6f0742f4d689 service nova] Acquiring lock "c2f2a485-363e-41af-9b3f-37092947ca08-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1190.640105] env[62692]: DEBUG oslo_concurrency.lockutils [req-4b1d3b55-ae3f-40f9-ad43-a8c1afb0e6bc req-67897676-6cc0-42e4-96b9-6f0742f4d689 service nova] Lock "c2f2a485-363e-41af-9b3f-37092947ca08-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1190.640287] env[62692]: DEBUG oslo_concurrency.lockutils [req-4b1d3b55-ae3f-40f9-ad43-a8c1afb0e6bc req-67897676-6cc0-42e4-96b9-6f0742f4d689 service nova] Lock "c2f2a485-363e-41af-9b3f-37092947ca08-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1190.640452] env[62692]: DEBUG nova.compute.manager [req-4b1d3b55-ae3f-40f9-ad43-a8c1afb0e6bc req-67897676-6cc0-42e4-96b9-6f0742f4d689 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] No waiting events found dispatching network-vif-unplugged-fa8d79b4-2bf8-4b6b-a5c2-6893f618f420 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1190.640653] env[62692]: WARNING nova.compute.manager [req-4b1d3b55-ae3f-40f9-ad43-a8c1afb0e6bc req-67897676-6cc0-42e4-96b9-6f0742f4d689 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Received unexpected event network-vif-unplugged-fa8d79b4-2bf8-4b6b-a5c2-6893f618f420 for instance with vm_state shelved and task_state shelving_offloading. [ 1190.656553] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142037, 'name': ReconfigVM_Task, 'duration_secs': 0.301017} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.659672] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Reconfigured VM instance instance-00000074 to attach disk [datastore1] 10d13b95-e6a1-4a11-9d8c-540ce82316de/10d13b95-e6a1-4a11-9d8c-540ce82316de.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1190.660561] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0f7de570-581d-44d6-bcbf-91aa86f6323f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.667582] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]525037fd-9020-6022-c90e-630f741a4e24, 'name': SearchDatastore_Task, 'duration_secs': 0.028602} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.668881] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1190.669210] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f/c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1190.669545] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1190.669545] env[62692]: value = "task-1142038" [ 1190.669545] env[62692]: _type = "Task" [ 1190.669545] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.671916] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b8d7aab6-c5fd-42b0-a538-0f4709928612 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.683467] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142038, 'name': Rename_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.684864] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1190.684864] env[62692]: value = "task-1142039" [ 1190.684864] env[62692]: _type = "Task" [ 1190.684864] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.695995] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142039, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.746347] env[62692]: DEBUG nova.network.neutron [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Updating instance_info_cache with network_info: [{"id": "21e76034-6c0f-4938-afda-0d005cf45759", "address": "fa:16:3e:2d:c0:4a", "network": {"id": "77695f27-3027-4a92-8b9b-0a90cdc53f62", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-337681783-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697b4281872f4547ad3d051e644b95ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21e76034-6c", "ovs_interfaceid": "21e76034-6c0f-4938-afda-0d005cf45759", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1190.759744] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1190.760879] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa8c9f8e-ffdd-4ea4-858e-7121a8e5038a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.769636] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1190.769899] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b9dcc65e-24e6-4ece-affd-db0b0107169a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.830876] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1190.831206] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7e7112e4-8b20-4d92-b5f4-8f9123524b16 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.839327] env[62692]: DEBUG oslo_vmware.api [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1190.839327] env[62692]: value = "task-1142041" [ 1190.839327] env[62692]: _type = "Task" [ 1190.839327] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.850249] env[62692]: DEBUG oslo_vmware.api [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1142041, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.852467] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1190.852688] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1190.852878] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Deleting the datastore file [datastore1] c2f2a485-363e-41af-9b3f-37092947ca08 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1190.853147] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-efa9df8a-94cb-4c98-8cb8-8c0cb126dc05 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.860325] env[62692]: DEBUG oslo_vmware.api [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1190.860325] env[62692]: value = "task-1142042" [ 1190.860325] env[62692]: _type = "Task" [ 1190.860325] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.868592] env[62692]: DEBUG oslo_vmware.api [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142042, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.054364] env[62692]: DEBUG oslo_concurrency.lockutils [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.785s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.082890] env[62692]: INFO nova.scheduler.client.report [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Deleted allocations for instance dd8a0d53-5dc8-4b66-add7-f244e600468d [ 1191.088824] env[62692]: INFO nova.compute.manager [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Took 17.08 seconds to build instance. [ 1191.141679] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 46bce865-188b-487c-a73f-cf524059c797] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1191.184819] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142038, 'name': Rename_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.196038] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142039, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.248966] env[62692]: DEBUG oslo_concurrency.lockutils [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Releasing lock "refresh_cache-90932f05-fad3-4c6d-87ae-ab059351b0be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1191.349084] env[62692]: DEBUG oslo_vmware.api [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1142041, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.370377] env[62692]: DEBUG oslo_vmware.api [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142042, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.381973] env[62692]: DEBUG oslo_concurrency.lockutils [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquiring lock "9d754b1a-0e1b-4200-804e-9e4fa3bd06bf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1191.592997] env[62692]: DEBUG oslo_concurrency.lockutils [None req-826005a6-0f3c-4e27-a220-e72494ad8697 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "9d754b1a-0e1b-4200-804e-9e4fa3bd06bf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.591s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.593513] env[62692]: DEBUG oslo_concurrency.lockutils [None req-742b2759-8f8a-4f34-8f54-48f7776b74a0 tempest-ServerRescueTestJSON-1332564777 tempest-ServerRescueTestJSON-1332564777-project-member] Lock "dd8a0d53-5dc8-4b66-add7-f244e600468d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.794s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.594388] env[62692]: DEBUG oslo_concurrency.lockutils [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "9d754b1a-0e1b-4200-804e-9e4fa3bd06bf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.213s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.594612] env[62692]: DEBUG oslo_concurrency.lockutils [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquiring lock "9d754b1a-0e1b-4200-804e-9e4fa3bd06bf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1191.594820] env[62692]: DEBUG oslo_concurrency.lockutils [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "9d754b1a-0e1b-4200-804e-9e4fa3bd06bf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.594988] env[62692]: DEBUG oslo_concurrency.lockutils [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "9d754b1a-0e1b-4200-804e-9e4fa3bd06bf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.596564] env[62692]: INFO nova.compute.manager [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Terminating instance [ 1191.598568] env[62692]: DEBUG nova.compute.manager [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1191.598568] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1191.599258] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-653559a3-00f9-4513-992e-a2d130472281 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.607329] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1191.607796] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7b437553-40bb-4304-b1dd-0b48863b0858 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.615603] env[62692]: DEBUG oslo_vmware.api [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1191.615603] env[62692]: value = "task-1142043" [ 1191.615603] env[62692]: _type = "Task" [ 1191.615603] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.627485] env[62692]: DEBUG oslo_vmware.api [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142043, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.645640] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: a668a569-a90e-4e59-8d2e-b0225745b500] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1191.685638] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142038, 'name': Rename_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.695701] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142039, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.681325} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.697075] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f/c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1191.697075] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1191.697075] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-72245d65-7e94-42ad-a0ad-b6cdac6819a0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.704669] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1191.704669] env[62692]: value = "task-1142044" [ 1191.704669] env[62692]: _type = "Task" [ 1191.704669] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.714759] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142044, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.850360] env[62692]: DEBUG oslo_vmware.api [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1142041, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.870696] env[62692]: DEBUG oslo_vmware.api [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142042, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.721023} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.870967] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1191.871179] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1191.871364] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1191.897748] env[62692]: INFO nova.scheduler.client.report [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Deleted allocations for instance c2f2a485-363e-41af-9b3f-37092947ca08 [ 1192.125992] env[62692]: DEBUG oslo_vmware.api [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142043, 'name': PowerOffVM_Task, 'duration_secs': 0.219786} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.126319] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1192.126644] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1192.126745] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3334b6fc-19f5-4b21-aaa6-d65c1740f1f0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.149307] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 3f6c9744-a6e3-43f3-8b6b-624ba178747d] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1192.185030] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142038, 'name': Rename_Task, 'duration_secs': 1.191422} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.186228] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1192.186556] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7702d12f-6afc-42aa-b458-ed1ccd8c93d3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.193979] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1192.194257] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1192.194488] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Deleting the datastore file [datastore1] 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1192.195815] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-85371d92-cdc1-4500-add2-be6f009513c1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.198039] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1192.198039] env[62692]: value = "task-1142046" [ 1192.198039] env[62692]: _type = "Task" [ 1192.198039] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.204309] env[62692]: DEBUG oslo_vmware.api [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1192.204309] env[62692]: value = "task-1142047" [ 1192.204309] env[62692]: _type = "Task" [ 1192.204309] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.207724] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142046, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.219204] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142044, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.102672} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.222559] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1192.222825] env[62692]: DEBUG oslo_vmware.api [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142047, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.223613] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d74510f7-1953-48a5-b2b6-8d92ecb3019a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.247038] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Reconfiguring VM instance instance-00000075 to attach disk [datastore1] c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f/c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1192.249290] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-66dac8e0-4ce5-4722-8dfa-a3313828b58e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.271374] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1192.271374] env[62692]: value = "task-1142048" [ 1192.271374] env[62692]: _type = "Task" [ 1192.271374] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.283439] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142048, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.350583] env[62692]: DEBUG oslo_vmware.api [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1142041, 'name': PowerOffVM_Task, 'duration_secs': 1.155215} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.350889] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1192.351095] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Updating instance '973ac4ab-5f82-4108-8076-9a370ca9d914' progress to 17 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1192.403498] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1192.403782] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1192.404018] env[62692]: DEBUG nova.objects.instance [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lazy-loading 'resources' on Instance uuid c2f2a485-363e-41af-9b3f-37092947ca08 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1192.654262] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: e2b3b062-10e2-4a9c-b6b3-f0786de07b79] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1192.669228] env[62692]: DEBUG nova.compute.manager [req-2a5dafa4-9638-4b72-a813-913901dd2412 req-d3cec2af-0b5b-4334-a24d-2a566488de94 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Received event network-changed-fa8d79b4-2bf8-4b6b-a5c2-6893f618f420 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1192.669228] env[62692]: DEBUG nova.compute.manager [req-2a5dafa4-9638-4b72-a813-913901dd2412 req-d3cec2af-0b5b-4334-a24d-2a566488de94 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Refreshing instance network info cache due to event network-changed-fa8d79b4-2bf8-4b6b-a5c2-6893f618f420. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1192.669228] env[62692]: DEBUG oslo_concurrency.lockutils [req-2a5dafa4-9638-4b72-a813-913901dd2412 req-d3cec2af-0b5b-4334-a24d-2a566488de94 service nova] Acquiring lock "refresh_cache-c2f2a485-363e-41af-9b3f-37092947ca08" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1192.669228] env[62692]: DEBUG oslo_concurrency.lockutils [req-2a5dafa4-9638-4b72-a813-913901dd2412 req-d3cec2af-0b5b-4334-a24d-2a566488de94 service nova] Acquired lock "refresh_cache-c2f2a485-363e-41af-9b3f-37092947ca08" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1192.669496] env[62692]: DEBUG nova.network.neutron [req-2a5dafa4-9638-4b72-a813-913901dd2412 req-d3cec2af-0b5b-4334-a24d-2a566488de94 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Refreshing network info cache for port fa8d79b4-2bf8-4b6b-a5c2-6893f618f420 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1192.708359] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142046, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.716985] env[62692]: DEBUG oslo_vmware.api [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142047, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170354} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.717254] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1192.717444] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1192.717699] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1192.717902] env[62692]: INFO nova.compute.manager [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1192.718203] env[62692]: DEBUG oslo.service.loopingcall [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1192.718404] env[62692]: DEBUG nova.compute.manager [-] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1192.718504] env[62692]: DEBUG nova.network.neutron [-] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1192.770234] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a03a341-5f9a-485e-af70-a90b22948ef8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.781344] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142048, 'name': ReconfigVM_Task, 'duration_secs': 0.395093} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.795772] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Reconfigured VM instance instance-00000075 to attach disk [datastore1] c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f/c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1192.796619] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Updating instance '90932f05-fad3-4c6d-87ae-ab059351b0be' progress to 0 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1192.799891] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b01a5776-4307-4327-95e7-cc5f0415c17c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.811214] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1192.811214] env[62692]: value = "task-1142049" [ 1192.811214] env[62692]: _type = "Task" [ 1192.811214] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.819756] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142049, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.857964] env[62692]: DEBUG nova.virt.hardware [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1192.858261] env[62692]: DEBUG nova.virt.hardware [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1192.858438] env[62692]: DEBUG nova.virt.hardware [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1192.858630] env[62692]: DEBUG nova.virt.hardware [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1192.858782] env[62692]: DEBUG nova.virt.hardware [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1192.858933] env[62692]: DEBUG nova.virt.hardware [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1192.859163] env[62692]: DEBUG nova.virt.hardware [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1192.859333] env[62692]: DEBUG nova.virt.hardware [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1192.859513] env[62692]: DEBUG nova.virt.hardware [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1192.859683] env[62692]: DEBUG nova.virt.hardware [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1192.859865] env[62692]: DEBUG nova.virt.hardware [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1192.865778] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68c91bbb-7041-43a6-abb8-d4faa54072e8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.883285] env[62692]: DEBUG oslo_vmware.api [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1192.883285] env[62692]: value = "task-1142050" [ 1192.883285] env[62692]: _type = "Task" [ 1192.883285] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.894244] env[62692]: DEBUG oslo_vmware.api [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1142050, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.908107] env[62692]: DEBUG nova.objects.instance [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lazy-loading 'numa_topology' on Instance uuid c2f2a485-363e-41af-9b3f-37092947ca08 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1193.158208] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 8a56e6df-eea1-41f4-9360-4f06d2f516a3] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1193.211133] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142046, 'name': PowerOnVM_Task, 'duration_secs': 0.534046} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.211415] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1193.211618] env[62692]: INFO nova.compute.manager [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Took 9.01 seconds to spawn the instance on the hypervisor. [ 1193.211857] env[62692]: DEBUG nova.compute.manager [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1193.212644] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ca082c3-eac8-42fb-b0d5-bdb19280b17d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.305024] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1193.305024] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-804150d0-8c5a-4fdb-812b-dd471be3724d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.317027] env[62692]: DEBUG oslo_vmware.api [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1193.317027] env[62692]: value = "task-1142051" [ 1193.317027] env[62692]: _type = "Task" [ 1193.317027] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.324123] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142049, 'name': Rename_Task, 'duration_secs': 0.265766} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.327033] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1193.327318] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8636bcf6-1859-42fd-8cc5-37bb2a7bab13 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.337074] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] VM already powered off {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1193.337074] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Updating instance '90932f05-fad3-4c6d-87ae-ab059351b0be' progress to 17 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1193.344082] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1193.344082] env[62692]: value = "task-1142052" [ 1193.344082] env[62692]: _type = "Task" [ 1193.344082] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.352018] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142052, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.395887] env[62692]: DEBUG oslo_vmware.api [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1142050, 'name': ReconfigVM_Task, 'duration_secs': 0.220531} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.397192] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Updating instance '973ac4ab-5f82-4108-8076-9a370ca9d914' progress to 33 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1193.412720] env[62692]: DEBUG nova.objects.base [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=62692) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1193.561912] env[62692]: DEBUG nova.network.neutron [req-2a5dafa4-9638-4b72-a813-913901dd2412 req-d3cec2af-0b5b-4334-a24d-2a566488de94 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Updated VIF entry in instance network info cache for port fa8d79b4-2bf8-4b6b-a5c2-6893f618f420. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1193.562340] env[62692]: DEBUG nova.network.neutron [req-2a5dafa4-9638-4b72-a813-913901dd2412 req-d3cec2af-0b5b-4334-a24d-2a566488de94 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Updating instance_info_cache with network_info: [{"id": "fa8d79b4-2bf8-4b6b-a5c2-6893f618f420", "address": "fa:16:3e:a0:b7:ee", "network": {"id": "22795f7f-18d7-4b10-8051-bcff4e0af3ac", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-445901728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.250", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96b921004f8a4bf68640c3810fe7b819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapfa8d79b4-2b", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1193.613256] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4db238cf-6772-4b19-a4f3-0ed430c5a800 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.621960] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4a55942-545b-49f0-9cc7-468498ecd473 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.655118] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de6505ce-fd42-4299-b032-f4d2d79e7565 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.662583] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: b23d3484-4fc5-46b2-8e8c-bdaab087f046] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1193.665379] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ad39a53-3e41-4b75-995d-b31025bc571c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.679432] env[62692]: DEBUG nova.compute.provider_tree [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1193.738708] env[62692]: INFO nova.compute.manager [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Took 18.98 seconds to build instance. [ 1193.843890] env[62692]: DEBUG nova.virt.hardware [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1193.844083] env[62692]: DEBUG nova.virt.hardware [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1193.844163] env[62692]: DEBUG nova.virt.hardware [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1193.844354] env[62692]: DEBUG nova.virt.hardware [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1193.844500] env[62692]: DEBUG nova.virt.hardware [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1193.844672] env[62692]: DEBUG nova.virt.hardware [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1193.844955] env[62692]: DEBUG nova.virt.hardware [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1193.845561] env[62692]: DEBUG nova.virt.hardware [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1193.845561] env[62692]: DEBUG nova.virt.hardware [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1193.845561] env[62692]: DEBUG nova.virt.hardware [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1193.845714] env[62692]: DEBUG nova.virt.hardware [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1193.851077] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5cb91fb8-ef26-4c9f-b4ea-00b88a724881 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.870467] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142052, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.871751] env[62692]: DEBUG oslo_vmware.api [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1193.871751] env[62692]: value = "task-1142053" [ 1193.871751] env[62692]: _type = "Task" [ 1193.871751] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.879997] env[62692]: DEBUG oslo_vmware.api [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142053, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.889855] env[62692]: DEBUG nova.network.neutron [-] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1193.907685] env[62692]: DEBUG nova.virt.hardware [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1193.907949] env[62692]: DEBUG nova.virt.hardware [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1193.908135] env[62692]: DEBUG nova.virt.hardware [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1193.908331] env[62692]: DEBUG nova.virt.hardware [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1193.908484] env[62692]: DEBUG nova.virt.hardware [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1193.908636] env[62692]: DEBUG nova.virt.hardware [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1193.908843] env[62692]: DEBUG nova.virt.hardware [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1193.909021] env[62692]: DEBUG nova.virt.hardware [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1193.909203] env[62692]: DEBUG nova.virt.hardware [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1193.909368] env[62692]: DEBUG nova.virt.hardware [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1193.909546] env[62692]: DEBUG nova.virt.hardware [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1194.065806] env[62692]: DEBUG oslo_concurrency.lockutils [req-2a5dafa4-9638-4b72-a813-913901dd2412 req-d3cec2af-0b5b-4334-a24d-2a566488de94 service nova] Releasing lock "refresh_cache-c2f2a485-363e-41af-9b3f-37092947ca08" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1194.170831] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 0ca10fa9-8bc2-43bb-a4cb-66bbcebbc13d] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1194.182499] env[62692]: DEBUG nova.scheduler.client.report [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1194.241326] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "10d13b95-e6a1-4a11-9d8c-540ce82316de" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.494s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1194.361541] env[62692]: DEBUG oslo_vmware.api [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142052, 'name': PowerOnVM_Task, 'duration_secs': 0.763338} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.361821] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1194.362038] env[62692]: INFO nova.compute.manager [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Took 7.84 seconds to spawn the instance on the hypervisor. [ 1194.362230] env[62692]: DEBUG nova.compute.manager [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1194.363084] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a954dc5e-938e-4fe9-84fb-f677d6fcd0d6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.380462] env[62692]: DEBUG oslo_vmware.api [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142053, 'name': ReconfigVM_Task, 'duration_secs': 0.149349} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.380748] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Updating instance '90932f05-fad3-4c6d-87ae-ab059351b0be' progress to 33 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1194.392529] env[62692]: INFO nova.compute.manager [-] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Took 1.67 seconds to deallocate network for instance. [ 1194.421030] env[62692]: ERROR nova.compute.manager [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Setting instance vm_state to ERROR: AttributeError: 'NoneType' object has no attribute 'key' [ 1194.421030] env[62692]: ERROR nova.compute.manager [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Traceback (most recent call last): [ 1194.421030] env[62692]: ERROR nova.compute.manager [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] File "/opt/stack/nova/nova/compute/manager.py", line 10863, in _error_out_instance_on_exception [ 1194.421030] env[62692]: ERROR nova.compute.manager [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] yield [ 1194.421030] env[62692]: ERROR nova.compute.manager [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] File "/opt/stack/nova/nova/compute/manager.py", line 6105, in _resize_instance [ 1194.421030] env[62692]: ERROR nova.compute.manager [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] disk_info = self.driver.migrate_disk_and_power_off( [ 1194.421030] env[62692]: ERROR nova.compute.manager [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 263, in migrate_disk_and_power_off [ 1194.421030] env[62692]: ERROR nova.compute.manager [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] return self._vmops.migrate_disk_and_power_off(context, instance, [ 1194.421030] env[62692]: ERROR nova.compute.manager [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1467, in migrate_disk_and_power_off [ 1194.421030] env[62692]: ERROR nova.compute.manager [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] self._resize_disk(instance, vm_ref, vmdk, flavor) [ 1194.421030] env[62692]: ERROR nova.compute.manager [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1398, in _resize_disk [ 1194.421030] env[62692]: ERROR nova.compute.manager [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] self._volumeops.detach_disk_from_vm(vm_ref, instance, vmdk.device) [ 1194.421030] env[62692]: ERROR nova.compute.manager [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 121, in detach_disk_from_vm [ 1194.421030] env[62692]: ERROR nova.compute.manager [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] disk_key = device.key [ 1194.421030] env[62692]: ERROR nova.compute.manager [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] AttributeError: 'NoneType' object has no attribute 'key' [ 1194.421030] env[62692]: ERROR nova.compute.manager [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] [ 1194.674425] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: fac426b6-747a-455e-85a1-44e71b432180] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1194.687502] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.284s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1194.773123] env[62692]: DEBUG nova.compute.manager [req-452691a0-24c8-4637-a566-14ccd0e8863f req-b130aa54-cbf7-473d-a851-7b0c258b53e0 service nova] [instance: 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf] Received event network-vif-deleted-9970a338-22fc-4fbf-9c8d-e2a446b21682 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1194.791061] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "c2f2a485-363e-41af-9b3f-37092947ca08" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1194.880317] env[62692]: INFO nova.compute.manager [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Took 20.08 seconds to build instance. [ 1194.886668] env[62692]: DEBUG nova.virt.hardware [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1194.886859] env[62692]: DEBUG nova.virt.hardware [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1194.887018] env[62692]: DEBUG nova.virt.hardware [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1194.887235] env[62692]: DEBUG nova.virt.hardware [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1194.887390] env[62692]: DEBUG nova.virt.hardware [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1194.887571] env[62692]: DEBUG nova.virt.hardware [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1194.887786] env[62692]: DEBUG nova.virt.hardware [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1194.887956] env[62692]: DEBUG nova.virt.hardware [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1194.888356] env[62692]: DEBUG nova.virt.hardware [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1194.888356] env[62692]: DEBUG nova.virt.hardware [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1194.888490] env[62692]: DEBUG nova.virt.hardware [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1194.894166] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Reconfiguring VM instance instance-0000005e to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1194.895022] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f004bbab-7990-4ada-8809-c3950c7b2a23 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.910240] env[62692]: DEBUG oslo_concurrency.lockutils [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1194.910483] env[62692]: DEBUG oslo_concurrency.lockutils [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1194.910712] env[62692]: DEBUG nova.objects.instance [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lazy-loading 'resources' on Instance uuid 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1194.920294] env[62692]: DEBUG oslo_vmware.api [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1194.920294] env[62692]: value = "task-1142054" [ 1194.920294] env[62692]: _type = "Task" [ 1194.920294] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.933117] env[62692]: DEBUG oslo_vmware.api [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142054, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.941253] env[62692]: INFO nova.compute.manager [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Swapping old allocation on dict_keys(['470758a8-5efa-4875-aad5-f512a727752f']) held by migration 0b6ea356-8353-433e-a19a-6a87dd74287a for instance [ 1194.966431] env[62692]: DEBUG nova.scheduler.client.report [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Overwriting current allocation {'allocations': {'470758a8-5efa-4875-aad5-f512a727752f': {'resources': {'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 159}}, 'project_id': '37921f67664544478cb2568def068b11', 'user_id': '016134ac36fd4bd893226ae95a955e28', 'consumer_generation': 1} on consumer 973ac4ab-5f82-4108-8076-9a370ca9d914 {{(pid=62692) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1195.177916] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: fcdd1a0e-3f20-4f56-9d9f-167c0e8c6957] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1195.196888] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ed869113-4e43-4f3c-9c60-53f8724e3b8a tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "c2f2a485-363e-41af-9b3f-37092947ca08" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 24.151s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1195.198159] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "c2f2a485-363e-41af-9b3f-37092947ca08" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.407s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1195.198490] env[62692]: INFO nova.compute.manager [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Unshelving [ 1195.342979] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "10d13b95-e6a1-4a11-9d8c-540ce82316de" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1195.343368] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "10d13b95-e6a1-4a11-9d8c-540ce82316de" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1195.344024] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "10d13b95-e6a1-4a11-9d8c-540ce82316de-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1195.344024] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "10d13b95-e6a1-4a11-9d8c-540ce82316de-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1195.344154] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "10d13b95-e6a1-4a11-9d8c-540ce82316de-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1195.346641] env[62692]: INFO nova.compute.manager [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Terminating instance [ 1195.348603] env[62692]: DEBUG nova.compute.manager [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1195.348810] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1195.349734] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f96875-1512-4583-9bde-7d60b8f932ad {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.358781] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1195.359134] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5da6a87e-5bd0-471c-beb1-f3a104640f4d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.366234] env[62692]: DEBUG oslo_vmware.api [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1195.366234] env[62692]: value = "task-1142055" [ 1195.366234] env[62692]: _type = "Task" [ 1195.366234] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.374875] env[62692]: DEBUG oslo_vmware.api [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142055, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.382678] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8edea6a3-3fa8-4e58-ab31-44a25f94c1d1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.598s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1195.417813] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1195.418129] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1195.418635] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1195.418635] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1195.418744] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1195.420823] env[62692]: INFO nova.compute.manager [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Terminating instance [ 1195.425666] env[62692]: DEBUG nova.compute.manager [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1195.425874] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1195.426661] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f4b7087-b2c7-4b9a-b332-f7969f77733b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.439073] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1195.439073] env[62692]: DEBUG oslo_vmware.api [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142054, 'name': ReconfigVM_Task, 'duration_secs': 0.244356} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.442252] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9cc12a21-f29d-46fe-b26c-b57164f7ce47 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "0b7ea87c-252a-48b3-a1cd-c81e91362812" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1195.442505] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9cc12a21-f29d-46fe-b26c-b57164f7ce47 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "0b7ea87c-252a-48b3-a1cd-c81e91362812" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1195.447018] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dbb950e7-4293-4ff6-8720-0fce78b5f216 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.447171] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Reconfigured VM instance instance-0000005e to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1195.451355] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37197870-fc66-4eee-a2b9-1a613e0ae160 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.477584] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 90932f05-fad3-4c6d-87ae-ab059351b0be/90932f05-fad3-4c6d-87ae-ab059351b0be.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1195.483318] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2267012c-ca96-4dc9-a9c5-ceae92d6e755 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.499077] env[62692]: DEBUG oslo_vmware.api [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1195.499077] env[62692]: value = "task-1142056" [ 1195.499077] env[62692]: _type = "Task" [ 1195.499077] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.511393] env[62692]: DEBUG oslo_vmware.api [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1195.511393] env[62692]: value = "task-1142057" [ 1195.511393] env[62692]: _type = "Task" [ 1195.511393] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.515323] env[62692]: DEBUG oslo_vmware.api [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142056, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.531703] env[62692]: DEBUG oslo_vmware.api [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142057, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.681820] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 44699df6-5262-4700-a096-8c0e450323d2] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1195.692386] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68fc58e0-1803-45c8-8d45-b31d1a82e4de {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.713043] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77a8eafa-1a68-4d71-9856-1fa2ec7b7db4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.746830] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e90826a-ecc0-4970-b4ff-139cb6fe124d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.757728] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-705f1da8-63aa-4665-ae20-933fcf61cd67 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.772640] env[62692]: DEBUG nova.compute.provider_tree [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1195.876315] env[62692]: DEBUG oslo_vmware.api [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142055, 'name': PowerOffVM_Task, 'duration_secs': 0.321368} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.876597] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1195.876813] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1195.877087] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-69af283c-379a-412d-b2c3-73dd671f628e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.943081] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1195.943297] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1195.943533] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Deleting the datastore file [datastore1] 10d13b95-e6a1-4a11-9d8c-540ce82316de {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1195.943840] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-efd9dcf3-787d-459f-b8b2-bed72fb2227b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.949077] env[62692]: DEBUG nova.compute.utils [None req-9cc12a21-f29d-46fe-b26c-b57164f7ce47 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1195.952808] env[62692]: DEBUG oslo_vmware.api [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1195.952808] env[62692]: value = "task-1142059" [ 1195.952808] env[62692]: _type = "Task" [ 1195.952808] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.962428] env[62692]: DEBUG oslo_vmware.api [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142059, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.004579] env[62692]: DEBUG oslo_concurrency.lockutils [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "973ac4ab-5f82-4108-8076-9a370ca9d914" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1196.004889] env[62692]: DEBUG oslo_concurrency.lockutils [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "973ac4ab-5f82-4108-8076-9a370ca9d914" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1196.005148] env[62692]: DEBUG oslo_concurrency.lockutils [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "973ac4ab-5f82-4108-8076-9a370ca9d914-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1196.005432] env[62692]: DEBUG oslo_concurrency.lockutils [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "973ac4ab-5f82-4108-8076-9a370ca9d914-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1196.005661] env[62692]: DEBUG oslo_concurrency.lockutils [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "973ac4ab-5f82-4108-8076-9a370ca9d914-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1196.011162] env[62692]: INFO nova.compute.manager [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Terminating instance [ 1196.013691] env[62692]: DEBUG nova.compute.manager [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1196.013945] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1196.014222] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-25d54f98-2087-495b-883d-32d0ad5dd5a3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.018924] env[62692]: DEBUG oslo_vmware.api [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142056, 'name': PowerOffVM_Task, 'duration_secs': 0.232021} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.022377] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1196.022590] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1196.022837] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8557bab7-b2bd-401b-a91d-9ec7bd28f1a9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.034995] env[62692]: DEBUG oslo_vmware.api [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1196.034995] env[62692]: value = "task-1142060" [ 1196.034995] env[62692]: _type = "Task" [ 1196.034995] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.035253] env[62692]: DEBUG oslo_vmware.api [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142057, 'name': ReconfigVM_Task, 'duration_secs': 0.356012} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.035791] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 90932f05-fad3-4c6d-87ae-ab059351b0be/90932f05-fad3-4c6d-87ae-ab059351b0be.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1196.036086] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Updating instance '90932f05-fad3-4c6d-87ae-ab059351b0be' progress to 50 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1196.048189] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] VM already powered off {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1196.048425] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Volume detach. Driver type: vmdk {{(pid=62692) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1196.048628] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249070', 'volume_id': 'd3635284-5881-414b-8d43-075170b6ad32', 'name': 'volume-d3635284-5881-414b-8d43-075170b6ad32', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '973ac4ab-5f82-4108-8076-9a370ca9d914', 'attached_at': '', 'detached_at': '', 'volume_id': 'd3635284-5881-414b-8d43-075170b6ad32', 'serial': 'd3635284-5881-414b-8d43-075170b6ad32'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1196.049407] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8612657f-a9d1-4b0d-9443-d412c189d4e6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.069803] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b8239a6-b6b3-42a2-83f6-65f328663272 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.078660] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-053f31bc-39e6-459d-a879-e37d366f4d3e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.101162] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fee91ee-8b01-4e45-acda-c8a70140ddeb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.106085] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1196.106304] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1196.106497] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Deleting the datastore file [datastore1] c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1196.106736] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9d3a3696-5a4b-4b81-8121-abfdca11a9b6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.121782] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] The volume has not been displaced from its original location: [datastore1] volume-d3635284-5881-414b-8d43-075170b6ad32/volume-d3635284-5881-414b-8d43-075170b6ad32.vmdk. No consolidation needed. {{(pid=62692) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1196.127075] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Reconfiguring VM instance instance-00000070 to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1196.128482] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-700c2be7-57f8-43bd-960a-75f334289131 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.141449] env[62692]: DEBUG oslo_vmware.api [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1196.141449] env[62692]: value = "task-1142062" [ 1196.141449] env[62692]: _type = "Task" [ 1196.141449] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.147335] env[62692]: DEBUG oslo_vmware.api [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1196.147335] env[62692]: value = "task-1142063" [ 1196.147335] env[62692]: _type = "Task" [ 1196.147335] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.152989] env[62692]: DEBUG oslo_vmware.api [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142062, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.158058] env[62692]: DEBUG oslo_vmware.api [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1142063, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.188106] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 3fd2959b-80cb-470c-8c2a-40b7630458ed] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1196.219810] env[62692]: DEBUG nova.compute.utils [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1196.275893] env[62692]: DEBUG nova.scheduler.client.report [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1196.455096] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9cc12a21-f29d-46fe-b26c-b57164f7ce47 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "0b7ea87c-252a-48b3-a1cd-c81e91362812" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.012s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1196.466025] env[62692]: DEBUG oslo_vmware.api [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142059, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187428} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.466428] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1196.466715] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1196.466952] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1196.467884] env[62692]: INFO nova.compute.manager [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1196.467884] env[62692]: DEBUG oslo.service.loopingcall [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1196.467884] env[62692]: DEBUG nova.compute.manager [-] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1196.467884] env[62692]: DEBUG nova.network.neutron [-] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1196.540224] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1196.545468] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d887c1a-0d09-446e-906c-c40032d82acb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.568326] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbb17bd6-c588-41cb-a344-61d585a47fba {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.589933] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Updating instance '90932f05-fad3-4c6d-87ae-ab059351b0be' progress to 67 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1196.654335] env[62692]: DEBUG oslo_vmware.api [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142062, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.227885} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.655270] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1196.655270] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1196.655532] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1196.655730] env[62692]: INFO nova.compute.manager [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Took 1.23 seconds to destroy the instance on the hypervisor. [ 1196.655971] env[62692]: DEBUG oslo.service.loopingcall [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1196.656554] env[62692]: DEBUG nova.compute.manager [-] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1196.656554] env[62692]: DEBUG nova.network.neutron [-] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1196.663088] env[62692]: DEBUG oslo_vmware.api [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1142063, 'name': ReconfigVM_Task, 'duration_secs': 0.198008} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.663811] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Reconfigured VM instance instance-00000070 to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1196.669230] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a9ec734-ddaf-4d59-9af4-dbbea96b6fff {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.685404] env[62692]: DEBUG oslo_vmware.api [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1196.685404] env[62692]: value = "task-1142064" [ 1196.685404] env[62692]: _type = "Task" [ 1196.685404] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.691355] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 9f86d7c7-79f7-46ea-bb02-fdcd1b8bc58e] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1196.696745] env[62692]: DEBUG oslo_vmware.api [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1142064, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.722834] env[62692]: INFO nova.virt.block_device [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Booting with volume c77c2006-709b-42d0-a47c-106f2a3f7237 at /dev/sdb [ 1196.761943] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bebf5eb8-4567-4260-b47d-93b32efa6afe {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.772717] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efc20155-e869-452e-92b3-009c17646554 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.783876] env[62692]: DEBUG oslo_concurrency.lockutils [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.873s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1196.786289] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.247s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1196.831021] env[62692]: INFO nova.scheduler.client.report [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Deleted allocations for instance 9d754b1a-0e1b-4200-804e-9e4fa3bd06bf [ 1196.831021] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6f711b65-3ba7-429d-8752-75941ad13dd7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.843177] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2767fe8d-9dea-47f9-b682-bda67dd090f8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.882485] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5ec978c-29f6-4883-8cba-acec6f799153 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.892249] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca376e34-5f9d-4270-9024-643ed0c51d46 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.896624] env[62692]: DEBUG nova.compute.manager [req-4ff5996d-ff59-4408-97f3-e5957de52b60 req-3e0c5444-4117-4010-8b06-22326e802383 service nova] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Received event network-vif-deleted-dff1785d-8de1-4070-9d59-44f56eb42854 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1196.896795] env[62692]: INFO nova.compute.manager [req-4ff5996d-ff59-4408-97f3-e5957de52b60 req-3e0c5444-4117-4010-8b06-22326e802383 service nova] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Neutron deleted interface dff1785d-8de1-4070-9d59-44f56eb42854; detaching it from the instance and deleting it from the info cache [ 1196.896970] env[62692]: DEBUG nova.network.neutron [req-4ff5996d-ff59-4408-97f3-e5957de52b60 req-3e0c5444-4117-4010-8b06-22326e802383 service nova] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1196.913181] env[62692]: DEBUG nova.virt.block_device [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Updating existing volume attachment record: b92a48fc-b907-4b2f-9a4e-f8b9d6d96819 {{(pid=62692) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1197.046770] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b592414-a148-40e2-beba-495e90ce7aea {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.058167] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6d33e42-b238-4913-aa75-c3f0090ffb8f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.094226] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc7796ec-4cb1-49a4-8202-ba298d86d0b7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.105388] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8db5d2d5-c2f9-479d-bfd9-d5658c0b00f5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.120440] env[62692]: DEBUG nova.compute.provider_tree [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1197.152181] env[62692]: DEBUG nova.network.neutron [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Port 21e76034-6c0f-4938-afda-0d005cf45759 binding to destination host cpu-1 is already ACTIVE {{(pid=62692) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1197.195889] env[62692]: DEBUG oslo_vmware.api [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1142064, 'name': ReconfigVM_Task, 'duration_secs': 0.152556} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.196264] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249070', 'volume_id': 'd3635284-5881-414b-8d43-075170b6ad32', 'name': 'volume-d3635284-5881-414b-8d43-075170b6ad32', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '973ac4ab-5f82-4108-8076-9a370ca9d914', 'attached_at': '', 'detached_at': '', 'volume_id': 'd3635284-5881-414b-8d43-075170b6ad32', 'serial': 'd3635284-5881-414b-8d43-075170b6ad32'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1197.196518] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1197.197619] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3bf08a3-560e-4066-8740-7694075781fb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.200391] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: ae44f074-a5c8-4259-99c6-9ce290b1570b] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1197.206421] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1197.206757] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f806ddc3-8b53-41d5-bd34-bcfe07703771 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.271606] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1197.271897] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1197.272225] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Deleting the datastore file [datastore1] 973ac4ab-5f82-4108-8076-9a370ca9d914 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1197.272586] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-40da2c05-819d-44b2-88c0-7701887ae3ab {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.280396] env[62692]: DEBUG oslo_vmware.api [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1197.280396] env[62692]: value = "task-1142069" [ 1197.280396] env[62692]: _type = "Task" [ 1197.280396] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.289516] env[62692]: DEBUG oslo_vmware.api [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1142069, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.310689] env[62692]: DEBUG nova.network.neutron [-] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1197.338764] env[62692]: DEBUG oslo_concurrency.lockutils [None req-80a1bc4c-6383-492e-9d20-c907e88134af tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "9d754b1a-0e1b-4200-804e-9e4fa3bd06bf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.744s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.403047] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-009a52a7-7fec-4724-b111-b4ce943bc7f8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.413210] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-051521d6-4546-4dd0-b1eb-48c9d6c3e32f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.441996] env[62692]: DEBUG nova.compute.manager [req-4ff5996d-ff59-4408-97f3-e5957de52b60 req-3e0c5444-4117-4010-8b06-22326e802383 service nova] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Detach interface failed, port_id=dff1785d-8de1-4070-9d59-44f56eb42854, reason: Instance 10d13b95-e6a1-4a11-9d8c-540ce82316de could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1197.474774] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ff49279a-358d-4747-a2ca-9c40afbbecc5 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "973ac4ab-5f82-4108-8076-9a370ca9d914" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.526113] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9cc12a21-f29d-46fe-b26c-b57164f7ce47 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "0b7ea87c-252a-48b3-a1cd-c81e91362812" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.526113] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9cc12a21-f29d-46fe-b26c-b57164f7ce47 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "0b7ea87c-252a-48b3-a1cd-c81e91362812" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.526113] env[62692]: INFO nova.compute.manager [None req-9cc12a21-f29d-46fe-b26c-b57164f7ce47 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Attaching volume 954f432b-2b81-4220-b6d2-9ab0ac462307 to /dev/sdb [ 1197.535292] env[62692]: DEBUG nova.network.neutron [-] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1197.556911] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cf3f39a-d863-469e-9a5a-c7a46171862e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.565053] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d428c85-8f5a-4ffd-bfe9-da39a0473632 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.578869] env[62692]: DEBUG nova.virt.block_device [None req-9cc12a21-f29d-46fe-b26c-b57164f7ce47 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Updating existing volume attachment record: b988c6ce-c3f5-40e7-9f37-fcb428fcc7b9 {{(pid=62692) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1197.623869] env[62692]: DEBUG nova.scheduler.client.report [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1197.703331] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 3dbe2a77-91aa-4de6-b9b1-1f30bd52ed19] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1197.790869] env[62692]: DEBUG oslo_vmware.api [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1142069, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.218544} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.791208] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1197.791455] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1197.791663] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1197.791857] env[62692]: INFO nova.compute.manager [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Took 1.78 seconds to destroy the instance on the hypervisor. [ 1197.792139] env[62692]: DEBUG oslo.service.loopingcall [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1197.792372] env[62692]: DEBUG nova.compute.manager [-] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1197.792466] env[62692]: DEBUG nova.network.neutron [-] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1197.813513] env[62692]: INFO nova.compute.manager [-] [instance: 10d13b95-e6a1-4a11-9d8c-540ce82316de] Took 1.35 seconds to deallocate network for instance. [ 1198.038381] env[62692]: INFO nova.compute.manager [-] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Took 1.38 seconds to deallocate network for instance. [ 1198.128469] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.342s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1198.128712] env[62692]: INFO nova.compute.manager [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Successfully reverted task state from resize_migrating on failure for instance. [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server [None req-9c2a5d89-6776-4f12-a68a-3b6a29a84abb tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Exception during message handling: AttributeError: 'NoneType' object has no attribute 'key' [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 172, in _process_incoming [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server raise self.value [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server raise self.value [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 213, in decorated_function [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server raise self.value [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6071, in resize_instance [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server raise self.value [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6068, in resize_instance [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server self._resize_instance(context, instance, image, migration, [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6105, in _resize_instance [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server disk_info = self.driver.migrate_disk_and_power_off( [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 263, in migrate_disk_and_power_off [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server return self._vmops.migrate_disk_and_power_off(context, instance, [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1467, in migrate_disk_and_power_off [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server self._resize_disk(instance, vm_ref, vmdk, flavor) [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1398, in _resize_disk [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server self._volumeops.detach_disk_from_vm(vm_ref, instance, vmdk.device) [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 121, in detach_disk_from_vm [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server disk_key = device.key [ 1198.140642] env[62692]: ERROR oslo_messaging.rpc.server AttributeError: 'NoneType' object has no attribute 'key' [ 1198.142662] env[62692]: ERROR oslo_messaging.rpc.server [ 1198.179434] env[62692]: DEBUG oslo_concurrency.lockutils [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "90932f05-fad3-4c6d-87ae-ab059351b0be-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1198.179748] env[62692]: DEBUG oslo_concurrency.lockutils [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "90932f05-fad3-4c6d-87ae-ab059351b0be-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1198.179997] env[62692]: DEBUG oslo_concurrency.lockutils [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "90932f05-fad3-4c6d-87ae-ab059351b0be-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1198.207567] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: b389b14d-6dcc-4b26-84f8-38b952bf6ecd] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1198.321153] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1198.321460] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1198.321708] env[62692]: DEBUG nova.objects.instance [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lazy-loading 'resources' on Instance uuid 10d13b95-e6a1-4a11-9d8c-540ce82316de {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1198.546816] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1198.710691] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 433f740c-bde4-4d33-baca-5a7ac8179da1] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1198.749329] env[62692]: DEBUG nova.network.neutron [-] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1198.929196] env[62692]: DEBUG nova.compute.manager [req-4ba1061d-18d9-4220-a4b2-4482bade722e req-e9bc52a6-0423-4c25-9f98-4166ac480290 service nova] [instance: c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f] Received event network-vif-deleted-fad50fb6-51f6-4917-beea-f67bf1334512 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1198.929416] env[62692]: DEBUG nova.compute.manager [req-4ba1061d-18d9-4220-a4b2-4482bade722e req-e9bc52a6-0423-4c25-9f98-4166ac480290 service nova] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Received event network-vif-deleted-e2bdb2a3-7e2c-4329-9008-9bea98b0b80a {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1198.947047] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "1c7ce408-6e9c-43bd-8d81-6acb12353ed8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1198.947154] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "1c7ce408-6e9c-43bd-8d81-6acb12353ed8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1198.947400] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "1c7ce408-6e9c-43bd-8d81-6acb12353ed8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1198.947629] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "1c7ce408-6e9c-43bd-8d81-6acb12353ed8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1198.947838] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "1c7ce408-6e9c-43bd-8d81-6acb12353ed8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1198.950014] env[62692]: INFO nova.compute.manager [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Terminating instance [ 1198.951885] env[62692]: DEBUG nova.compute.manager [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1198.952101] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1198.955205] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d0cf5db-727f-4744-b165-a3f103d7046b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.963122] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1198.963377] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8134d1b2-4729-4f2c-8e56-7f1e9d060fc1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.969269] env[62692]: DEBUG oslo_vmware.api [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1198.969269] env[62692]: value = "task-1142071" [ 1198.969269] env[62692]: _type = "Task" [ 1198.969269] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.979134] env[62692]: DEBUG oslo_vmware.api [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1142071, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.981695] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30d4ef86-97d8-4c99-945a-ad3ac49cd472 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.988355] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38b9f41a-b0cd-403b-baef-f37c81da6c97 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.019779] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87126952-c4dd-4e54-ba0b-25c6604c6f6e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.028054] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65dfd496-d31e-4ca4-bfae-ccdee8679dc1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.041654] env[62692]: DEBUG nova.compute.provider_tree [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1199.214066] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: f44e2ee2-6321-4237-9644-a19b5e90ae16] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1199.217275] env[62692]: DEBUG oslo_concurrency.lockutils [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "refresh_cache-90932f05-fad3-4c6d-87ae-ab059351b0be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1199.217449] env[62692]: DEBUG oslo_concurrency.lockutils [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquired lock "refresh_cache-90932f05-fad3-4c6d-87ae-ab059351b0be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1199.217626] env[62692]: DEBUG nova.network.neutron [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1199.252099] env[62692]: INFO nova.compute.manager [-] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Took 1.46 seconds to deallocate network for instance. [ 1199.481683] env[62692]: DEBUG oslo_vmware.api [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1142071, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.544427] env[62692]: DEBUG nova.scheduler.client.report [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1199.717536] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 26c1ff5b-9eda-4131-a4ea-d4511a2364b7] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1199.953217] env[62692]: DEBUG nova.network.neutron [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Updating instance_info_cache with network_info: [{"id": "21e76034-6c0f-4938-afda-0d005cf45759", "address": "fa:16:3e:2d:c0:4a", "network": {"id": "77695f27-3027-4a92-8b9b-0a90cdc53f62", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-337681783-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697b4281872f4547ad3d051e644b95ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21e76034-6c", "ovs_interfaceid": "21e76034-6c0f-4938-afda-0d005cf45759", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1199.982491] env[62692]: DEBUG oslo_vmware.api [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1142071, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.050087] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.728s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1200.052198] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.505s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1200.052622] env[62692]: DEBUG nova.objects.instance [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lazy-loading 'resources' on Instance uuid c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1200.072663] env[62692]: INFO nova.scheduler.client.report [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Deleted allocations for instance 10d13b95-e6a1-4a11-9d8c-540ce82316de [ 1200.221310] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 754b3c65-1e4b-49d2-8980-095d975edb01] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1200.456263] env[62692]: DEBUG oslo_concurrency.lockutils [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Releasing lock "refresh_cache-90932f05-fad3-4c6d-87ae-ab059351b0be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1200.482597] env[62692]: DEBUG oslo_vmware.api [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1142071, 'name': PowerOffVM_Task, 'duration_secs': 1.132421} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.482864] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1200.483182] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1200.483310] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ab0aff76-a8ef-45d5-b96f-5785eddfce1f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.549023] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1200.549268] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1200.549494] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Deleting the datastore file [datastore2] 1c7ce408-6e9c-43bd-8d81-6acb12353ed8 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1200.549766] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6d3765c8-c5b6-425a-9f4a-162f761eca6d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.558962] env[62692]: DEBUG oslo_vmware.api [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1200.558962] env[62692]: value = "task-1142075" [ 1200.558962] env[62692]: _type = "Task" [ 1200.558962] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.567656] env[62692]: DEBUG oslo_vmware.api [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1142075, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.579303] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ba5ab9c5-4d86-4259-9a51-2f34e04c8ca9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "10d13b95-e6a1-4a11-9d8c-540ce82316de" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.236s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1200.701185] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae169725-5386-4dad-8570-82f235387d47 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.709096] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03997733-2af8-470c-93e3-f77141ad91dc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.738409] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 6b2d3385-a9af-432b-a1e7-c91adb2e1f69] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1200.741065] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2421418-c6b2-4902-9477-3b583d15ef3c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.748893] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9529c02a-ea1b-4864-9895-d4c957c9dff7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.762645] env[62692]: DEBUG nova.compute.provider_tree [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1200.983043] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a7e9d30-9fe0-43c2-9a2b-7e0e1b79129c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.001665] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e216cbf3-9419-497b-8ed9-2701675851ee {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.009655] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Updating instance '90932f05-fad3-4c6d-87ae-ab059351b0be' progress to 83 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1201.068537] env[62692]: DEBUG oslo_vmware.api [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1142075, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136583} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.068804] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1201.068996] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1201.069193] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1201.069372] env[62692]: INFO nova.compute.manager [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Took 2.12 seconds to destroy the instance on the hypervisor. [ 1201.069659] env[62692]: DEBUG oslo.service.loopingcall [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1201.069887] env[62692]: DEBUG nova.compute.manager [-] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1201.069963] env[62692]: DEBUG nova.network.neutron [-] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1201.244761] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 18eb004f-a16b-477a-8d05-ed23a418b069] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1201.266475] env[62692]: DEBUG nova.scheduler.client.report [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1201.368320] env[62692]: DEBUG nova.compute.manager [req-2366d8af-73f3-434b-a7c5-bc96ebe5e44f req-cf5bce8f-5f77-4a7f-8419-c2da28207d52 service nova] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Received event network-vif-deleted-5c943de3-e4d0-479b-9d1d-b38816b891b2 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1201.368515] env[62692]: INFO nova.compute.manager [req-2366d8af-73f3-434b-a7c5-bc96ebe5e44f req-cf5bce8f-5f77-4a7f-8419-c2da28207d52 service nova] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Neutron deleted interface 5c943de3-e4d0-479b-9d1d-b38816b891b2; detaching it from the instance and deleting it from the info cache [ 1201.368674] env[62692]: DEBUG nova.network.neutron [req-2366d8af-73f3-434b-a7c5-bc96ebe5e44f req-cf5bce8f-5f77-4a7f-8419-c2da28207d52 service nova] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1201.515772] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-074853ac-d629-4ca7-94b4-96d0341d59d4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Updating instance '90932f05-fad3-4c6d-87ae-ab059351b0be' progress to 100 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1201.561904] env[62692]: DEBUG oslo_concurrency.lockutils [None req-31d255b4-d57f-4f3e-9591-93acb7b85e5a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "45eccffc-f059-4bef-97da-dacb16f20f88" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1201.562159] env[62692]: DEBUG oslo_concurrency.lockutils [None req-31d255b4-d57f-4f3e-9591-93acb7b85e5a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "45eccffc-f059-4bef-97da-dacb16f20f88" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1201.748309] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 64a8329d-e0e9-4c2d-bd1f-844ee40eb980] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1201.771662] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.719s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1201.796820] env[62692]: INFO nova.scheduler.client.report [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Deleted allocations for instance c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f [ 1201.848728] env[62692]: DEBUG nova.network.neutron [-] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1201.871086] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f3338691-7a7b-473b-8b6f-c8e2ae891068 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.881079] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b74b914b-2aaf-4768-ad29-71a76c5a8955 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.910846] env[62692]: DEBUG nova.compute.manager [req-2366d8af-73f3-434b-a7c5-bc96ebe5e44f req-cf5bce8f-5f77-4a7f-8419-c2da28207d52 service nova] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Detach interface failed, port_id=5c943de3-e4d0-479b-9d1d-b38816b891b2, reason: Instance 1c7ce408-6e9c-43bd-8d81-6acb12353ed8 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1202.065127] env[62692]: DEBUG nova.compute.utils [None req-31d255b4-d57f-4f3e-9591-93acb7b85e5a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1202.121779] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-9cc12a21-f29d-46fe-b26c-b57164f7ce47 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Volume attach. Driver type: vmdk {{(pid=62692) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1202.122041] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-9cc12a21-f29d-46fe-b26c-b57164f7ce47 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249090', 'volume_id': '954f432b-2b81-4220-b6d2-9ab0ac462307', 'name': 'volume-954f432b-2b81-4220-b6d2-9ab0ac462307', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0b7ea87c-252a-48b3-a1cd-c81e91362812', 'attached_at': '', 'detached_at': '', 'volume_id': '954f432b-2b81-4220-b6d2-9ab0ac462307', 'serial': '954f432b-2b81-4220-b6d2-9ab0ac462307'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1202.123043] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5ddd1b6-4831-4c98-a5cd-dbca1aac031c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.140512] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd355fdf-d3c6-42f6-8faa-c3a9551c2877 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.166481] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-9cc12a21-f29d-46fe-b26c-b57164f7ce47 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] volume-954f432b-2b81-4220-b6d2-9ab0ac462307/volume-954f432b-2b81-4220-b6d2-9ab0ac462307.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1202.168616] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-98e1f4da-b20b-4710-b438-b3336848e11d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.188830] env[62692]: DEBUG oslo_vmware.api [None req-9cc12a21-f29d-46fe-b26c-b57164f7ce47 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1202.188830] env[62692]: value = "task-1142076" [ 1202.188830] env[62692]: _type = "Task" [ 1202.188830] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.197604] env[62692]: DEBUG oslo_vmware.api [None req-9cc12a21-f29d-46fe-b26c-b57164f7ce47 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1142076, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.234669] env[62692]: INFO nova.compute.manager [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Took 2.98 seconds to detach 1 volumes for instance. [ 1202.238940] env[62692]: DEBUG nova.compute.manager [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Deleting volume: d3635284-5881-414b-8d43-075170b6ad32 {{(pid=62692) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1202.251772] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: c6a965d3-1c52-4f95-a226-9d15b7197ce7] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1202.306922] env[62692]: DEBUG oslo_concurrency.lockutils [None req-8a99857e-0073-409d-8087-d2fdc37f48b1 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "c7a8cb2f-bd4e-4499-9286-aa00fc79dc8f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.889s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1202.351384] env[62692]: INFO nova.compute.manager [-] [instance: 1c7ce408-6e9c-43bd-8d81-6acb12353ed8] Took 1.28 seconds to deallocate network for instance. [ 1202.569011] env[62692]: DEBUG oslo_concurrency.lockutils [None req-31d255b4-d57f-4f3e-9591-93acb7b85e5a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "45eccffc-f059-4bef-97da-dacb16f20f88" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1202.701377] env[62692]: DEBUG oslo_vmware.api [None req-9cc12a21-f29d-46fe-b26c-b57164f7ce47 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1142076, 'name': ReconfigVM_Task, 'duration_secs': 0.380173} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.702709] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1202.702957] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1202.703188] env[62692]: DEBUG nova.objects.instance [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lazy-loading 'pci_requests' on Instance uuid c2f2a485-363e-41af-9b3f-37092947ca08 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1202.704263] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-9cc12a21-f29d-46fe-b26c-b57164f7ce47 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Reconfigured VM instance instance-0000006e to attach disk [datastore2] volume-954f432b-2b81-4220-b6d2-9ab0ac462307/volume-954f432b-2b81-4220-b6d2-9ab0ac462307.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1202.709779] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-92807b4c-250f-4170-9682-657227ee0620 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.726594] env[62692]: DEBUG oslo_vmware.api [None req-9cc12a21-f29d-46fe-b26c-b57164f7ce47 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1202.726594] env[62692]: value = "task-1142078" [ 1202.726594] env[62692]: _type = "Task" [ 1202.726594] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.735105] env[62692]: DEBUG oslo_vmware.api [None req-9cc12a21-f29d-46fe-b26c-b57164f7ce47 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1142078, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.755079] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 2ccb76cf-f641-4306-a137-fb2417285df9] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1202.782390] env[62692]: DEBUG oslo_concurrency.lockutils [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1202.858425] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.211771] env[62692]: DEBUG nova.objects.instance [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lazy-loading 'numa_topology' on Instance uuid c2f2a485-363e-41af-9b3f-37092947ca08 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1203.237243] env[62692]: DEBUG oslo_vmware.api [None req-9cc12a21-f29d-46fe-b26c-b57164f7ce47 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1142078, 'name': ReconfigVM_Task, 'duration_secs': 0.145554} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.237624] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-9cc12a21-f29d-46fe-b26c-b57164f7ce47 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249090', 'volume_id': '954f432b-2b81-4220-b6d2-9ab0ac462307', 'name': 'volume-954f432b-2b81-4220-b6d2-9ab0ac462307', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0b7ea87c-252a-48b3-a1cd-c81e91362812', 'attached_at': '', 'detached_at': '', 'volume_id': '954f432b-2b81-4220-b6d2-9ab0ac462307', 'serial': '954f432b-2b81-4220-b6d2-9ab0ac462307'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1203.259980] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 81ce340c-fbef-4932-983c-595843530dbc] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1203.633554] env[62692]: DEBUG oslo_concurrency.lockutils [None req-31d255b4-d57f-4f3e-9591-93acb7b85e5a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "45eccffc-f059-4bef-97da-dacb16f20f88" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.633901] env[62692]: DEBUG oslo_concurrency.lockutils [None req-31d255b4-d57f-4f3e-9591-93acb7b85e5a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "45eccffc-f059-4bef-97da-dacb16f20f88" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1203.634108] env[62692]: INFO nova.compute.manager [None req-31d255b4-d57f-4f3e-9591-93acb7b85e5a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Attaching volume 89113b2e-696c-4171-8d9f-479a6c1f41db to /dev/sdb [ 1203.663296] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf5c2054-5c1a-4f9b-95c6-89b8d7bdf903 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.670857] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b49bc6b8-c8c6-4871-8188-912abd446d53 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.685197] env[62692]: DEBUG nova.virt.block_device [None req-31d255b4-d57f-4f3e-9591-93acb7b85e5a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Updating existing volume attachment record: 2a6507e7-35c8-4b1a-90f5-f909bb5f480e {{(pid=62692) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1203.715106] env[62692]: INFO nova.compute.claims [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1203.764544] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 27583cb4-0478-44d8-8970-478682271513] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1203.853751] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a52630f8-0a43-4cf1-9a0b-538bd76b3b85 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "90932f05-fad3-4c6d-87ae-ab059351b0be" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.856400] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a52630f8-0a43-4cf1-9a0b-538bd76b3b85 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "90932f05-fad3-4c6d-87ae-ab059351b0be" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1203.856400] env[62692]: DEBUG nova.compute.manager [None req-a52630f8-0a43-4cf1-9a0b-538bd76b3b85 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Going to confirm migration 8 {{(pid=62692) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1203.869476] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "29910848-22f2-4c2e-98fe-cd9e32e97865" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.869721] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "29910848-22f2-4c2e-98fe-cd9e32e97865" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1203.920144] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "0806e8c3-8b51-45e8-aaf8-cbd6f03f5489" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.920540] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "0806e8c3-8b51-45e8-aaf8-cbd6f03f5489" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1204.268251] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: ed9b23dc-1fbd-4a0e-a2fd-4a47b946ab64] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1204.276386] env[62692]: DEBUG nova.objects.instance [None req-9cc12a21-f29d-46fe-b26c-b57164f7ce47 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lazy-loading 'flavor' on Instance uuid 0b7ea87c-252a-48b3-a1cd-c81e91362812 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1204.376186] env[62692]: DEBUG nova.compute.manager [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1204.402545] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a52630f8-0a43-4cf1-9a0b-538bd76b3b85 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "refresh_cache-90932f05-fad3-4c6d-87ae-ab059351b0be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1204.402632] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a52630f8-0a43-4cf1-9a0b-538bd76b3b85 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquired lock "refresh_cache-90932f05-fad3-4c6d-87ae-ab059351b0be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1204.402812] env[62692]: DEBUG nova.network.neutron [None req-a52630f8-0a43-4cf1-9a0b-538bd76b3b85 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1204.403011] env[62692]: DEBUG nova.objects.instance [None req-a52630f8-0a43-4cf1-9a0b-538bd76b3b85 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lazy-loading 'info_cache' on Instance uuid 90932f05-fad3-4c6d-87ae-ab059351b0be {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1204.422970] env[62692]: DEBUG nova.compute.manager [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1204.771812] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: bb547773-d176-4c8e-a0fa-a374d5050b1b] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1204.784442] env[62692]: DEBUG oslo_concurrency.lockutils [None req-9cc12a21-f29d-46fe-b26c-b57164f7ce47 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "0b7ea87c-252a-48b3-a1cd-c81e91362812" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.259s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1204.895604] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8355e9b2-b9c5-4570-b998-5af41bd12492 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.898790] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1204.904290] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-863e5fa0-4fba-48a9-9cb7-ffb350e02f98 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.941800] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-661e8d85-afc9-435a-962a-55042aada785 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.950959] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c1a2fca-9343-427c-a03d-457ca17c44b1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.955821] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1204.964975] env[62692]: DEBUG nova.compute.provider_tree [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1205.275483] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 9b8e830c-61b7-4dd2-8324-d3a96eec1465] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1205.346033] env[62692]: DEBUG oslo_concurrency.lockutils [None req-091e524a-7f26-4690-8edf-472028386ef9 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "0b7ea87c-252a-48b3-a1cd-c81e91362812" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1205.346346] env[62692]: DEBUG oslo_concurrency.lockutils [None req-091e524a-7f26-4690-8edf-472028386ef9 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "0b7ea87c-252a-48b3-a1cd-c81e91362812" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1205.467678] env[62692]: DEBUG nova.scheduler.client.report [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1205.750281] env[62692]: DEBUG nova.network.neutron [None req-a52630f8-0a43-4cf1-9a0b-538bd76b3b85 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Updating instance_info_cache with network_info: [{"id": "21e76034-6c0f-4938-afda-0d005cf45759", "address": "fa:16:3e:2d:c0:4a", "network": {"id": "77695f27-3027-4a92-8b9b-0a90cdc53f62", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-337681783-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697b4281872f4547ad3d051e644b95ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21e76034-6c", "ovs_interfaceid": "21e76034-6c0f-4938-afda-0d005cf45759", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1205.779075] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 5a08fa8a-f9fe-4879-bb7b-baa04097df6c] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1205.849698] env[62692]: INFO nova.compute.manager [None req-091e524a-7f26-4690-8edf-472028386ef9 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Detaching volume 954f432b-2b81-4220-b6d2-9ab0ac462307 [ 1205.885326] env[62692]: INFO nova.virt.block_device [None req-091e524a-7f26-4690-8edf-472028386ef9 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Attempting to driver detach volume 954f432b-2b81-4220-b6d2-9ab0ac462307 from mountpoint /dev/sdb [ 1205.885678] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-091e524a-7f26-4690-8edf-472028386ef9 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Volume detach. Driver type: vmdk {{(pid=62692) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1205.885927] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-091e524a-7f26-4690-8edf-472028386ef9 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249090', 'volume_id': '954f432b-2b81-4220-b6d2-9ab0ac462307', 'name': 'volume-954f432b-2b81-4220-b6d2-9ab0ac462307', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0b7ea87c-252a-48b3-a1cd-c81e91362812', 'attached_at': '', 'detached_at': '', 'volume_id': '954f432b-2b81-4220-b6d2-9ab0ac462307', 'serial': '954f432b-2b81-4220-b6d2-9ab0ac462307'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1205.886882] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8592e49e-0a67-4a15-8795-bded66bad4fb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.910526] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e1a85ea-3dea-4db0-babe-16d047f366e2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.917942] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d80c10cb-4d57-49bc-8be9-b291835ede43 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.939133] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f16a176a-1214-451b-922b-ae6c0b0f640d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.955322] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-091e524a-7f26-4690-8edf-472028386ef9 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] The volume has not been displaced from its original location: [datastore2] volume-954f432b-2b81-4220-b6d2-9ab0ac462307/volume-954f432b-2b81-4220-b6d2-9ab0ac462307.vmdk. No consolidation needed. {{(pid=62692) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1205.961774] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-091e524a-7f26-4690-8edf-472028386ef9 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Reconfiguring VM instance instance-0000006e to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1205.962086] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3373af1e-e030-4c92-8192-0644c68605f7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.975219] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.272s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1205.977387] env[62692]: DEBUG oslo_concurrency.lockutils [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.195s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1205.977658] env[62692]: DEBUG nova.objects.instance [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lazy-loading 'resources' on Instance uuid 973ac4ab-5f82-4108-8076-9a370ca9d914 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1205.984845] env[62692]: DEBUG oslo_vmware.api [None req-091e524a-7f26-4690-8edf-472028386ef9 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1205.984845] env[62692]: value = "task-1142081" [ 1205.984845] env[62692]: _type = "Task" [ 1205.984845] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.994744] env[62692]: DEBUG oslo_vmware.api [None req-091e524a-7f26-4690-8edf-472028386ef9 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1142081, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.007446] env[62692]: INFO nova.network.neutron [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Updating port fa8d79b4-2bf8-4b6b-a5c2-6893f618f420 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1206.254663] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a52630f8-0a43-4cf1-9a0b-538bd76b3b85 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Releasing lock "refresh_cache-90932f05-fad3-4c6d-87ae-ab059351b0be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1206.254965] env[62692]: DEBUG nova.objects.instance [None req-a52630f8-0a43-4cf1-9a0b-538bd76b3b85 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lazy-loading 'migration_context' on Instance uuid 90932f05-fad3-4c6d-87ae-ab059351b0be {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1206.282555] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: abee88ff-5cf5-4bf1-91e0-93b19cf30046] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1206.495766] env[62692]: DEBUG oslo_vmware.api [None req-091e524a-7f26-4690-8edf-472028386ef9 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1142081, 'name': ReconfigVM_Task, 'duration_secs': 0.217823} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.496064] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-091e524a-7f26-4690-8edf-472028386ef9 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Reconfigured VM instance instance-0000006e to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1206.503247] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c71407d3-d8b5-4450-ad6e-73318e2c3360 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.519461] env[62692]: DEBUG oslo_vmware.api [None req-091e524a-7f26-4690-8edf-472028386ef9 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1206.519461] env[62692]: value = "task-1142082" [ 1206.519461] env[62692]: _type = "Task" [ 1206.519461] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.530412] env[62692]: DEBUG oslo_vmware.api [None req-091e524a-7f26-4690-8edf-472028386ef9 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1142082, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.650263] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d51b539-1876-4bed-ba03-0627a3858cd6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.658157] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed6534a7-daf9-4dfd-9e25-5e5aa7ee66bd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.689155] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a341cd2-689f-490d-b971-dc032ef2e73b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.696696] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cb54212-ccb2-4c1d-83ad-b26422693663 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.709620] env[62692]: DEBUG nova.compute.provider_tree [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1206.757875] env[62692]: DEBUG nova.objects.base [None req-a52630f8-0a43-4cf1-9a0b-538bd76b3b85 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Object Instance<90932f05-fad3-4c6d-87ae-ab059351b0be> lazy-loaded attributes: info_cache,migration_context {{(pid=62692) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1206.758840] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17ff6d02-6f8a-482c-8c7f-22459448889c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.778338] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ec26d3e-d419-454b-b85e-d16daba0bb81 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.784083] env[62692]: DEBUG oslo_vmware.api [None req-a52630f8-0a43-4cf1-9a0b-538bd76b3b85 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1206.784083] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]523b6e83-f251-c3dc-b958-5c7c4e1ea6c9" [ 1206.784083] env[62692]: _type = "Task" [ 1206.784083] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.787147] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 628cc505-3edf-4066-91be-da009ebcf219] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1206.802666] env[62692]: DEBUG oslo_vmware.api [None req-a52630f8-0a43-4cf1-9a0b-538bd76b3b85 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]523b6e83-f251-c3dc-b958-5c7c4e1ea6c9, 'name': SearchDatastore_Task, 'duration_secs': 0.015407} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.802853] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a52630f8-0a43-4cf1-9a0b-538bd76b3b85 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1207.029612] env[62692]: DEBUG oslo_vmware.api [None req-091e524a-7f26-4690-8edf-472028386ef9 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1142082, 'name': ReconfigVM_Task, 'duration_secs': 0.152707} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.029939] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-091e524a-7f26-4690-8edf-472028386ef9 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249090', 'volume_id': '954f432b-2b81-4220-b6d2-9ab0ac462307', 'name': 'volume-954f432b-2b81-4220-b6d2-9ab0ac462307', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0b7ea87c-252a-48b3-a1cd-c81e91362812', 'attached_at': '', 'detached_at': '', 'volume_id': '954f432b-2b81-4220-b6d2-9ab0ac462307', 'serial': '954f432b-2b81-4220-b6d2-9ab0ac462307'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1207.212969] env[62692]: DEBUG nova.scheduler.client.report [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1207.290234] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: d99737d8-2eb0-40ee-b61e-6c736c84ea59] Instance has had 0 of 5 cleanup attempts {{(pid=62692) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1207.549407] env[62692]: DEBUG nova.compute.manager [req-38a94b78-4dae-4645-b878-ddc92e1df96a req-d0a2835c-b84c-4bb3-8d70-801a82a8ff32 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Received event network-vif-plugged-fa8d79b4-2bf8-4b6b-a5c2-6893f618f420 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1207.549631] env[62692]: DEBUG oslo_concurrency.lockutils [req-38a94b78-4dae-4645-b878-ddc92e1df96a req-d0a2835c-b84c-4bb3-8d70-801a82a8ff32 service nova] Acquiring lock "c2f2a485-363e-41af-9b3f-37092947ca08-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1207.549993] env[62692]: DEBUG oslo_concurrency.lockutils [req-38a94b78-4dae-4645-b878-ddc92e1df96a req-d0a2835c-b84c-4bb3-8d70-801a82a8ff32 service nova] Lock "c2f2a485-363e-41af-9b3f-37092947ca08-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1207.549993] env[62692]: DEBUG oslo_concurrency.lockutils [req-38a94b78-4dae-4645-b878-ddc92e1df96a req-d0a2835c-b84c-4bb3-8d70-801a82a8ff32 service nova] Lock "c2f2a485-363e-41af-9b3f-37092947ca08-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1207.552285] env[62692]: DEBUG nova.compute.manager [req-38a94b78-4dae-4645-b878-ddc92e1df96a req-d0a2835c-b84c-4bb3-8d70-801a82a8ff32 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] No waiting events found dispatching network-vif-plugged-fa8d79b4-2bf8-4b6b-a5c2-6893f618f420 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1207.552511] env[62692]: WARNING nova.compute.manager [req-38a94b78-4dae-4645-b878-ddc92e1df96a req-d0a2835c-b84c-4bb3-8d70-801a82a8ff32 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Received unexpected event network-vif-plugged-fa8d79b4-2bf8-4b6b-a5c2-6893f618f420 for instance with vm_state shelved_offloaded and task_state spawning. [ 1207.579427] env[62692]: DEBUG nova.objects.instance [None req-091e524a-7f26-4690-8edf-472028386ef9 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lazy-loading 'flavor' on Instance uuid 0b7ea87c-252a-48b3-a1cd-c81e91362812 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1207.699591] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "refresh_cache-c2f2a485-363e-41af-9b3f-37092947ca08" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1207.699818] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquired lock "refresh_cache-c2f2a485-363e-41af-9b3f-37092947ca08" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1207.704826] env[62692]: DEBUG nova.network.neutron [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1207.721153] env[62692]: DEBUG oslo_concurrency.lockutils [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.744s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1207.723878] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.866s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1207.724362] env[62692]: DEBUG nova.objects.instance [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lazy-loading 'resources' on Instance uuid 1c7ce408-6e9c-43bd-8d81-6acb12353ed8 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1207.744197] env[62692]: INFO nova.scheduler.client.report [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Deleted allocations for instance 973ac4ab-5f82-4108-8076-9a370ca9d914 [ 1207.793899] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1207.794087] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Cleaning up deleted instances with incomplete migration {{(pid=62692) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 1208.232794] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-31d255b4-d57f-4f3e-9591-93acb7b85e5a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Volume attach. Driver type: vmdk {{(pid=62692) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1208.233731] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-31d255b4-d57f-4f3e-9591-93acb7b85e5a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249091', 'volume_id': '89113b2e-696c-4171-8d9f-479a6c1f41db', 'name': 'volume-89113b2e-696c-4171-8d9f-479a6c1f41db', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '45eccffc-f059-4bef-97da-dacb16f20f88', 'attached_at': '', 'detached_at': '', 'volume_id': '89113b2e-696c-4171-8d9f-479a6c1f41db', 'serial': '89113b2e-696c-4171-8d9f-479a6c1f41db'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1208.234746] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55be4f8b-5ada-4dbf-a72d-15be3b53460b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.257091] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ded2654d-843b-4180-85dc-a95a2e465fbf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.259967] env[62692]: DEBUG oslo_concurrency.lockutils [None req-75d101c9-8a28-4343-bd6b-82164a20e0f3 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "973ac4ab-5f82-4108-8076-9a370ca9d914" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.255s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1208.261575] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ff49279a-358d-4747-a2ca-9c40afbbecc5 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "973ac4ab-5f82-4108-8076-9a370ca9d914" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 10.787s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1208.261575] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ff49279a-358d-4747-a2ca-9c40afbbecc5 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "973ac4ab-5f82-4108-8076-9a370ca9d914-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1208.261575] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ff49279a-358d-4747-a2ca-9c40afbbecc5 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "973ac4ab-5f82-4108-8076-9a370ca9d914-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1208.262345] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ff49279a-358d-4747-a2ca-9c40afbbecc5 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "973ac4ab-5f82-4108-8076-9a370ca9d914-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1208.266569] env[62692]: INFO nova.compute.manager [None req-ff49279a-358d-4747-a2ca-9c40afbbecc5 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Terminating instance [ 1208.268581] env[62692]: DEBUG nova.compute.manager [None req-ff49279a-358d-4747-a2ca-9c40afbbecc5 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1208.268865] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-308604e1-ea5f-4448-979f-067d111a068a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.295503] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-31d255b4-d57f-4f3e-9591-93acb7b85e5a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] volume-89113b2e-696c-4171-8d9f-479a6c1f41db/volume-89113b2e-696c-4171-8d9f-479a6c1f41db.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1208.297500] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-959cd652-bb49-41fb-b99a-1333491dc2ad {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.316678] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fced5dc-b618-4edc-b3c1-6c9a5c25a673 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.330242] env[62692]: DEBUG oslo_vmware.api [None req-31d255b4-d57f-4f3e-9591-93acb7b85e5a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1208.330242] env[62692]: value = "task-1142083" [ 1208.330242] env[62692]: _type = "Task" [ 1208.330242] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.340971] env[62692]: DEBUG oslo_vmware.api [None req-31d255b4-d57f-4f3e-9591-93acb7b85e5a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1142083, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.352570] env[62692]: WARNING nova.virt.vmwareapi.driver [None req-ff49279a-358d-4747-a2ca-9c40afbbecc5 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 973ac4ab-5f82-4108-8076-9a370ca9d914 could not be found. [ 1208.352813] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ff49279a-358d-4747-a2ca-9c40afbbecc5 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1208.353596] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-91793dcd-e3bd-4ab6-ab7f-642d8426b8a5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.362136] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4ff9e63-16f7-4db8-8186-9cb0dd3ac706 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.393312] env[62692]: WARNING nova.virt.vmwareapi.vmops [None req-ff49279a-358d-4747-a2ca-9c40afbbecc5 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 973ac4ab-5f82-4108-8076-9a370ca9d914 could not be found. [ 1208.393545] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ff49279a-358d-4747-a2ca-9c40afbbecc5 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1208.393966] env[62692]: INFO nova.compute.manager [None req-ff49279a-358d-4747-a2ca-9c40afbbecc5 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Took 0.12 seconds to destroy the instance on the hypervisor. [ 1208.394297] env[62692]: DEBUG oslo.service.loopingcall [None req-ff49279a-358d-4747-a2ca-9c40afbbecc5 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1208.398691] env[62692]: DEBUG nova.compute.manager [-] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1208.398800] env[62692]: DEBUG nova.network.neutron [-] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1208.474195] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f429e37-86cf-4cf6-8cd8-caed71eff790 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.484608] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e45ad5fd-26e9-4c1a-ad81-d2bf86461809 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.517353] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a9fcd17-9776-47f1-8882-cd47c0d165e7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.525034] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68c7102f-46ab-4e5f-9c3c-8d1243e08c55 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.541042] env[62692]: DEBUG nova.compute.provider_tree [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1208.582725] env[62692]: DEBUG nova.network.neutron [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Updating instance_info_cache with network_info: [{"id": "fa8d79b4-2bf8-4b6b-a5c2-6893f618f420", "address": "fa:16:3e:a0:b7:ee", "network": {"id": "22795f7f-18d7-4b10-8051-bcff4e0af3ac", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-445901728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.250", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96b921004f8a4bf68640c3810fe7b819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa8d79b4-2b", "ovs_interfaceid": "fa8d79b4-2bf8-4b6b-a5c2-6893f618f420", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1208.590095] env[62692]: DEBUG oslo_concurrency.lockutils [None req-091e524a-7f26-4690-8edf-472028386ef9 tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "0b7ea87c-252a-48b3-a1cd-c81e91362812" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.244s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1208.841104] env[62692]: DEBUG oslo_vmware.api [None req-31d255b4-d57f-4f3e-9591-93acb7b85e5a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1142083, 'name': ReconfigVM_Task, 'duration_secs': 0.383841} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.841418] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-31d255b4-d57f-4f3e-9591-93acb7b85e5a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Reconfigured VM instance instance-0000006f to attach disk [datastore2] volume-89113b2e-696c-4171-8d9f-479a6c1f41db/volume-89113b2e-696c-4171-8d9f-479a6c1f41db.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1208.846384] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-644cb7b1-94ca-405b-97d6-40022684da22 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.861676] env[62692]: DEBUG oslo_vmware.api [None req-31d255b4-d57f-4f3e-9591-93acb7b85e5a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1208.861676] env[62692]: value = "task-1142084" [ 1208.861676] env[62692]: _type = "Task" [ 1208.861676] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.869806] env[62692]: DEBUG oslo_vmware.api [None req-31d255b4-d57f-4f3e-9591-93acb7b85e5a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1142084, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.923431] env[62692]: DEBUG nova.network.neutron [-] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1209.041467] env[62692]: DEBUG nova.scheduler.client.report [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1209.085016] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Releasing lock "refresh_cache-c2f2a485-363e-41af-9b3f-37092947ca08" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1209.111514] env[62692]: DEBUG nova.virt.hardware [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='dd4c27f05f90dd4d28fa2263d019fddd',container_format='bare',created_at=2024-10-05T20:39:15Z,direct_url=,disk_format='vmdk',id=c0e34614-043a-4a57-969f-c9e09d45f20e,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-984238047-shelved',owner='96b921004f8a4bf68640c3810fe7b819',properties=ImageMetaProps,protected=,size=31668736,status='active',tags=,updated_at=2024-10-05T20:39:31Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1209.111776] env[62692]: DEBUG nova.virt.hardware [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1209.111941] env[62692]: DEBUG nova.virt.hardware [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1209.112148] env[62692]: DEBUG nova.virt.hardware [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1209.112304] env[62692]: DEBUG nova.virt.hardware [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1209.112458] env[62692]: DEBUG nova.virt.hardware [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1209.112755] env[62692]: DEBUG nova.virt.hardware [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1209.112931] env[62692]: DEBUG nova.virt.hardware [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1209.113120] env[62692]: DEBUG nova.virt.hardware [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1209.113291] env[62692]: DEBUG nova.virt.hardware [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1209.113471] env[62692]: DEBUG nova.virt.hardware [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1209.114393] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ecc0ae2-564a-4d97-83da-60b0d3baf1a4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.122604] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c769600f-98f1-4274-b084-fedfba4c13e3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.136720] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a0:b7:ee', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd1da5fc2-0280-4f76-ac97-20ea4bc7bb16', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fa8d79b4-2bf8-4b6b-a5c2-6893f618f420', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1209.144537] env[62692]: DEBUG oslo.service.loopingcall [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1209.145799] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1209.148154] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f04ec87a-aeb6-449a-8dd0-a5d0dfc37415 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.170011] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1209.170011] env[62692]: value = "task-1142085" [ 1209.170011] env[62692]: _type = "Task" [ 1209.170011] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.178234] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1142085, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.372212] env[62692]: DEBUG oslo_vmware.api [None req-31d255b4-d57f-4f3e-9591-93acb7b85e5a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1142084, 'name': ReconfigVM_Task, 'duration_secs': 0.141879} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.372581] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-31d255b4-d57f-4f3e-9591-93acb7b85e5a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249091', 'volume_id': '89113b2e-696c-4171-8d9f-479a6c1f41db', 'name': 'volume-89113b2e-696c-4171-8d9f-479a6c1f41db', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '45eccffc-f059-4bef-97da-dacb16f20f88', 'attached_at': '', 'detached_at': '', 'volume_id': '89113b2e-696c-4171-8d9f-479a6c1f41db', 'serial': '89113b2e-696c-4171-8d9f-479a6c1f41db'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1209.426361] env[62692]: INFO nova.compute.manager [-] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Took 1.03 seconds to deallocate network for instance. [ 1209.441298] env[62692]: WARNING nova.volume.cinder [None req-ff49279a-358d-4747-a2ca-9c40afbbecc5 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Attachment e9bd0fc1-d683-4faa-a377-fb15a3882af9 does not exist. Ignoring.: cinderclient.exceptions.NotFound: Volume attachment could not be found with filter: attachment_id = e9bd0fc1-d683-4faa-a377-fb15a3882af9. (HTTP 404) (Request-ID: req-85592d01-2623-49b8-accc-c5c90208d155) [ 1209.441591] env[62692]: INFO nova.compute.manager [None req-ff49279a-358d-4747-a2ca-9c40afbbecc5 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Took 0.01 seconds to detach 1 volumes for instance. [ 1209.443950] env[62692]: DEBUG nova.compute.manager [None req-ff49279a-358d-4747-a2ca-9c40afbbecc5 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Deleting volume: d3635284-5881-414b-8d43-075170b6ad32 {{(pid=62692) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1209.458398] env[62692]: WARNING nova.compute.manager [None req-ff49279a-358d-4747-a2ca-9c40afbbecc5 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Failed to delete volume: d3635284-5881-414b-8d43-075170b6ad32 due to Volume d3635284-5881-414b-8d43-075170b6ad32 could not be found.: nova.exception.VolumeNotFound: Volume d3635284-5881-414b-8d43-075170b6ad32 could not be found. [ 1209.546710] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.823s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1209.549120] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.650s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1209.550650] env[62692]: INFO nova.compute.claims [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1209.571300] env[62692]: INFO nova.scheduler.client.report [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Deleted allocations for instance 1c7ce408-6e9c-43bd-8d81-6acb12353ed8 [ 1209.616127] env[62692]: DEBUG nova.compute.manager [req-b0dab1a1-8462-45cf-9376-007c2886d880 req-48f0484e-25ca-4310-8c3c-64a6650687d8 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Received event network-changed-fa8d79b4-2bf8-4b6b-a5c2-6893f618f420 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1209.616290] env[62692]: DEBUG nova.compute.manager [req-b0dab1a1-8462-45cf-9376-007c2886d880 req-48f0484e-25ca-4310-8c3c-64a6650687d8 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Refreshing instance network info cache due to event network-changed-fa8d79b4-2bf8-4b6b-a5c2-6893f618f420. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1209.616517] env[62692]: DEBUG oslo_concurrency.lockutils [req-b0dab1a1-8462-45cf-9376-007c2886d880 req-48f0484e-25ca-4310-8c3c-64a6650687d8 service nova] Acquiring lock "refresh_cache-c2f2a485-363e-41af-9b3f-37092947ca08" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1209.616662] env[62692]: DEBUG oslo_concurrency.lockutils [req-b0dab1a1-8462-45cf-9376-007c2886d880 req-48f0484e-25ca-4310-8c3c-64a6650687d8 service nova] Acquired lock "refresh_cache-c2f2a485-363e-41af-9b3f-37092947ca08" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1209.616828] env[62692]: DEBUG nova.network.neutron [req-b0dab1a1-8462-45cf-9376-007c2886d880 req-48f0484e-25ca-4310-8c3c-64a6650687d8 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Refreshing network info cache for port fa8d79b4-2bf8-4b6b-a5c2-6893f618f420 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1209.664380] env[62692]: DEBUG oslo_concurrency.lockutils [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "0b7ea87c-252a-48b3-a1cd-c81e91362812" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1209.664619] env[62692]: DEBUG oslo_concurrency.lockutils [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "0b7ea87c-252a-48b3-a1cd-c81e91362812" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1209.664974] env[62692]: DEBUG oslo_concurrency.lockutils [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "0b7ea87c-252a-48b3-a1cd-c81e91362812-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1209.665275] env[62692]: DEBUG oslo_concurrency.lockutils [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "0b7ea87c-252a-48b3-a1cd-c81e91362812-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1209.665477] env[62692]: DEBUG oslo_concurrency.lockutils [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "0b7ea87c-252a-48b3-a1cd-c81e91362812-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1209.667686] env[62692]: INFO nova.compute.manager [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Terminating instance [ 1209.675016] env[62692]: DEBUG nova.compute.manager [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1209.675232] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1209.675956] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e570a28-7718-4122-b6d5-15331bc0e0b3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.684256] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1142085, 'name': CreateVM_Task, 'duration_secs': 0.384838} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.686071] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1209.686354] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1209.686946] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0e34614-043a-4a57-969f-c9e09d45f20e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1209.687120] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0e34614-043a-4a57-969f-c9e09d45f20e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1209.687481] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0e34614-043a-4a57-969f-c9e09d45f20e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1209.687700] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ef5ff5db-6e65-41d3-9c29-c91d413054b6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.689025] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51bee7b0-6161-4295-8f0f-24fd2d9c869a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.693195] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1209.693195] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]523db2e2-53b3-3cd6-2cc7-8293d36fecfc" [ 1209.693195] env[62692]: _type = "Task" [ 1209.693195] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.697227] env[62692]: DEBUG oslo_vmware.api [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1209.697227] env[62692]: value = "task-1142086" [ 1209.697227] env[62692]: _type = "Task" [ 1209.697227] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.702968] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]523db2e2-53b3-3cd6-2cc7-8293d36fecfc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.709541] env[62692]: DEBUG oslo_vmware.api [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1142086, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.962851] env[62692]: INFO nova.compute.manager [None req-ff49279a-358d-4747-a2ca-9c40afbbecc5 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: 973ac4ab-5f82-4108-8076-9a370ca9d914] Instance disappeared during terminate [ 1209.963123] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ff49279a-358d-4747-a2ca-9c40afbbecc5 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "973ac4ab-5f82-4108-8076-9a370ca9d914" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 1.702s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1210.080483] env[62692]: DEBUG oslo_concurrency.lockutils [None req-4d1060ab-db49-4fa1-ac30-54446a0fae12 tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "1c7ce408-6e9c-43bd-8d81-6acb12353ed8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.133s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1210.204522] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0e34614-043a-4a57-969f-c9e09d45f20e" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1210.204828] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Processing image c0e34614-043a-4a57-969f-c9e09d45f20e {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1210.205129] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0e34614-043a-4a57-969f-c9e09d45f20e/c0e34614-043a-4a57-969f-c9e09d45f20e.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1210.205342] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0e34614-043a-4a57-969f-c9e09d45f20e/c0e34614-043a-4a57-969f-c9e09d45f20e.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1210.205615] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1210.208745] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-98d40c25-26ab-4da5-ae8f-2edbeb6d73ca {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.210477] env[62692]: DEBUG oslo_vmware.api [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1142086, 'name': PowerOffVM_Task, 'duration_secs': 0.187313} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.212782] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1210.212958] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1210.213527] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3741905a-ecd6-4af4-b74d-0b6f7c1e1f54 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.220455] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1210.220455] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1210.220775] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca47b4cc-3013-47ae-9b3b-74f2b18d40b8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.227662] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1210.227662] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5277e83f-2aa4-ab2b-bacf-6fc8bfdf7b55" [ 1210.227662] env[62692]: _type = "Task" [ 1210.227662] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.238506] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5277e83f-2aa4-ab2b-bacf-6fc8bfdf7b55, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.299727] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1210.299900] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1210.300096] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Deleting the datastore file [datastore2] 0b7ea87c-252a-48b3-a1cd-c81e91362812 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1210.300385] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7432f5c1-9766-426f-b23e-7c330a1b165f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.306240] env[62692]: DEBUG oslo_vmware.api [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for the task: (returnval){ [ 1210.306240] env[62692]: value = "task-1142088" [ 1210.306240] env[62692]: _type = "Task" [ 1210.306240] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.314099] env[62692]: DEBUG oslo_vmware.api [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1142088, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.425323] env[62692]: DEBUG nova.objects.instance [None req-31d255b4-d57f-4f3e-9591-93acb7b85e5a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lazy-loading 'flavor' on Instance uuid 45eccffc-f059-4bef-97da-dacb16f20f88 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1210.642747] env[62692]: DEBUG nova.network.neutron [req-b0dab1a1-8462-45cf-9376-007c2886d880 req-48f0484e-25ca-4310-8c3c-64a6650687d8 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Updated VIF entry in instance network info cache for port fa8d79b4-2bf8-4b6b-a5c2-6893f618f420. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1210.643217] env[62692]: DEBUG nova.network.neutron [req-b0dab1a1-8462-45cf-9376-007c2886d880 req-48f0484e-25ca-4310-8c3c-64a6650687d8 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Updating instance_info_cache with network_info: [{"id": "fa8d79b4-2bf8-4b6b-a5c2-6893f618f420", "address": "fa:16:3e:a0:b7:ee", "network": {"id": "22795f7f-18d7-4b10-8051-bcff4e0af3ac", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-445901728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.250", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96b921004f8a4bf68640c3810fe7b819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa8d79b4-2b", "ovs_interfaceid": "fa8d79b4-2bf8-4b6b-a5c2-6893f618f420", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1210.700782] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d978d3d0-481e-40cf-a8fd-c6cea0f31988 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.710404] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87c7ee1f-a60c-479c-aae8-7cabf36e50dc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.745027] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e541b30-bcdf-4566-aebe-d796b4ae3a8e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.755208] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Preparing fetch location {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1210.755480] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Fetch image to [datastore1] OSTACK_IMG_cad11812-0190-4a3b-a482-e1b2a143692e/OSTACK_IMG_cad11812-0190-4a3b-a482-e1b2a143692e.vmdk {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1210.755671] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Downloading stream optimized image c0e34614-043a-4a57-969f-c9e09d45f20e to [datastore1] OSTACK_IMG_cad11812-0190-4a3b-a482-e1b2a143692e/OSTACK_IMG_cad11812-0190-4a3b-a482-e1b2a143692e.vmdk on the data store datastore1 as vApp {{(pid=62692) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1210.755847] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Downloading image file data c0e34614-043a-4a57-969f-c9e09d45f20e to the ESX as VM named 'OSTACK_IMG_cad11812-0190-4a3b-a482-e1b2a143692e' {{(pid=62692) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1210.758871] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-797c559e-5ace-45a4-9903-7b84f1d9f356 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.772875] env[62692]: DEBUG nova.compute.provider_tree [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1210.819086] env[62692]: DEBUG oslo_vmware.api [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Task: {'id': task-1142088, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147237} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.819260] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1210.819449] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1210.819696] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1210.819826] env[62692]: INFO nova.compute.manager [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1210.820089] env[62692]: DEBUG oslo.service.loopingcall [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1210.820294] env[62692]: DEBUG nova.compute.manager [-] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1210.820390] env[62692]: DEBUG nova.network.neutron [-] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1210.833458] env[62692]: DEBUG oslo_vmware.rw_handles [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1210.833458] env[62692]: value = "resgroup-9" [ 1210.833458] env[62692]: _type = "ResourcePool" [ 1210.833458] env[62692]: }. {{(pid=62692) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1210.833817] env[62692]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-d0a6da53-9855-447d-bda2-abcf8f8b5d4b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.855253] env[62692]: DEBUG oslo_vmware.rw_handles [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lease: (returnval){ [ 1210.855253] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5237e562-bced-9ec0-a905-25008c936416" [ 1210.855253] env[62692]: _type = "HttpNfcLease" [ 1210.855253] env[62692]: } obtained for vApp import into resource pool (val){ [ 1210.855253] env[62692]: value = "resgroup-9" [ 1210.855253] env[62692]: _type = "ResourcePool" [ 1210.855253] env[62692]: }. {{(pid=62692) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1210.855632] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the lease: (returnval){ [ 1210.855632] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5237e562-bced-9ec0-a905-25008c936416" [ 1210.855632] env[62692]: _type = "HttpNfcLease" [ 1210.855632] env[62692]: } to be ready. {{(pid=62692) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1210.861577] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1210.861577] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5237e562-bced-9ec0-a905-25008c936416" [ 1210.861577] env[62692]: _type = "HttpNfcLease" [ 1210.861577] env[62692]: } is initializing. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1210.931288] env[62692]: DEBUG oslo_concurrency.lockutils [None req-31d255b4-d57f-4f3e-9591-93acb7b85e5a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "45eccffc-f059-4bef-97da-dacb16f20f88" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.297s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1211.148690] env[62692]: DEBUG oslo_concurrency.lockutils [req-b0dab1a1-8462-45cf-9376-007c2886d880 req-48f0484e-25ca-4310-8c3c-64a6650687d8 service nova] Releasing lock "refresh_cache-c2f2a485-363e-41af-9b3f-37092947ca08" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1211.277856] env[62692]: DEBUG nova.scheduler.client.report [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1211.366838] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1211.366838] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5237e562-bced-9ec0-a905-25008c936416" [ 1211.366838] env[62692]: _type = "HttpNfcLease" [ 1211.366838] env[62692]: } is initializing. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1211.639959] env[62692]: DEBUG nova.compute.manager [req-f80429d9-f7b6-4cb1-a120-18799f4a17af req-581fc91c-2140-4641-a0ab-86467bf8320c service nova] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Received event network-vif-deleted-34de5e92-0920-4a35-bb28-51c7d09bf35d {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1211.640239] env[62692]: INFO nova.compute.manager [req-f80429d9-f7b6-4cb1-a120-18799f4a17af req-581fc91c-2140-4641-a0ab-86467bf8320c service nova] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Neutron deleted interface 34de5e92-0920-4a35-bb28-51c7d09bf35d; detaching it from the instance and deleting it from the info cache [ 1211.640239] env[62692]: DEBUG nova.network.neutron [req-f80429d9-f7b6-4cb1-a120-18799f4a17af req-581fc91c-2140-4641-a0ab-86467bf8320c service nova] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1211.782149] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.233s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1211.782713] env[62692]: DEBUG nova.compute.manager [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1211.785220] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.829s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1211.786938] env[62692]: INFO nova.compute.claims [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1211.866050] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1211.866050] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5237e562-bced-9ec0-a905-25008c936416" [ 1211.866050] env[62692]: _type = "HttpNfcLease" [ 1211.866050] env[62692]: } is ready. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1211.866375] env[62692]: DEBUG oslo_vmware.rw_handles [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1211.866375] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5237e562-bced-9ec0-a905-25008c936416" [ 1211.866375] env[62692]: _type = "HttpNfcLease" [ 1211.866375] env[62692]: }. {{(pid=62692) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1211.867097] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e24e581-01d6-423e-9185-816840927cc6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.874985] env[62692]: DEBUG oslo_vmware.rw_handles [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522a128c-d944-0e78-52f3-a52d9ea0cea7/disk-0.vmdk from lease info. {{(pid=62692) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1211.875205] env[62692]: DEBUG oslo_vmware.rw_handles [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Creating HTTP connection to write to file with size = 31668736 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522a128c-d944-0e78-52f3-a52d9ea0cea7/disk-0.vmdk. {{(pid=62692) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1211.938941] env[62692]: DEBUG nova.network.neutron [-] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1211.946455] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9dc59d69-af8a-4e14-8042-f5c0f775552f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.088466] env[62692]: DEBUG oslo_concurrency.lockutils [None req-917ddd32-8ccb-4465-85b9-50e7f944d2c6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "45eccffc-f059-4bef-97da-dacb16f20f88" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1212.088663] env[62692]: DEBUG oslo_concurrency.lockutils [None req-917ddd32-8ccb-4465-85b9-50e7f944d2c6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "45eccffc-f059-4bef-97da-dacb16f20f88" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1212.144043] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4f47e4b4-46b8-4359-b873-f9ecabe7c9e4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.153821] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d07b1a4c-ada0-4f78-98d9-2adec233ef2f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.185872] env[62692]: DEBUG nova.compute.manager [req-f80429d9-f7b6-4cb1-a120-18799f4a17af req-581fc91c-2140-4641-a0ab-86467bf8320c service nova] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Detach interface failed, port_id=34de5e92-0920-4a35-bb28-51c7d09bf35d, reason: Instance 0b7ea87c-252a-48b3-a1cd-c81e91362812 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1212.290859] env[62692]: DEBUG nova.compute.utils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1212.295851] env[62692]: DEBUG nova.compute.manager [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1212.296041] env[62692]: DEBUG nova.network.neutron [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1212.352614] env[62692]: DEBUG nova.policy [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b7d10dd39f064a23ae4fa8a8ec777d92', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5a1ab0259e314d22b8f36ba6a9f4d693', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1212.439288] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquiring lock "8abacc6f-054f-4c4f-abb2-83f20748cf1c" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1212.439817] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "8abacc6f-054f-4c4f-abb2-83f20748cf1c" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1212.440154] env[62692]: INFO nova.compute.manager [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Shelving [ 1212.442618] env[62692]: INFO nova.compute.manager [-] [instance: 0b7ea87c-252a-48b3-a1cd-c81e91362812] Took 1.62 seconds to deallocate network for instance. [ 1212.592296] env[62692]: DEBUG nova.compute.utils [None req-917ddd32-8ccb-4465-85b9-50e7f944d2c6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1212.681510] env[62692]: DEBUG nova.network.neutron [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Successfully created port: e7e22660-7340-4f0b-8d51-ad4b8834c59b {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1212.799508] env[62692]: DEBUG nova.compute.manager [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1212.952455] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1212.952813] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ef126e11-ef96-4dd9-90c2-d886d804e013 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.957290] env[62692]: DEBUG oslo_concurrency.lockutils [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1212.975913] env[62692]: DEBUG oslo_vmware.api [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1212.975913] env[62692]: value = "task-1142091" [ 1212.975913] env[62692]: _type = "Task" [ 1212.975913] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.987301] env[62692]: DEBUG oslo_vmware.api [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142091, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.991310] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bb9d676-ec4d-4245-a223-c8eb140dd715 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.998538] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1e71c45-e397-45be-87d3-61576cb54f2e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.035561] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8f795cf-09e5-4a83-b51d-469faf6b98ec {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.044637] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6e57a82-7c05-4438-8e2f-e20554939529 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.064061] env[62692]: DEBUG nova.compute.provider_tree [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1213.098033] env[62692]: DEBUG oslo_concurrency.lockutils [None req-917ddd32-8ccb-4465-85b9-50e7f944d2c6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "45eccffc-f059-4bef-97da-dacb16f20f88" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1213.110452] env[62692]: DEBUG oslo_vmware.rw_handles [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Completed reading data from the image iterator. {{(pid=62692) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1213.110648] env[62692]: DEBUG oslo_vmware.rw_handles [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522a128c-d944-0e78-52f3-a52d9ea0cea7/disk-0.vmdk. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1213.111586] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd59d086-932a-4359-a2da-12a59f905e04 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.118644] env[62692]: DEBUG oslo_vmware.rw_handles [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522a128c-d944-0e78-52f3-a52d9ea0cea7/disk-0.vmdk is in state: ready. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1213.118813] env[62692]: DEBUG oslo_vmware.rw_handles [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522a128c-d944-0e78-52f3-a52d9ea0cea7/disk-0.vmdk. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1213.119114] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-844a7de5-8be5-4c68-9b1a-4c1353de2a2c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.423792] env[62692]: DEBUG oslo_vmware.rw_handles [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522a128c-d944-0e78-52f3-a52d9ea0cea7/disk-0.vmdk. {{(pid=62692) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1213.424075] env[62692]: INFO nova.virt.vmwareapi.images [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Downloaded image file data c0e34614-043a-4a57-969f-c9e09d45f20e [ 1213.425061] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81a789db-eb4a-420b-a023-8a6be7b26c03 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.440681] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-751e8ed2-5286-45e9-a057-c9160b9e03ed {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.482694] env[62692]: INFO nova.virt.vmwareapi.images [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] The imported VM was unregistered [ 1213.485980] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Caching image {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1213.486293] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Creating directory with path [datastore1] devstack-image-cache_base/c0e34614-043a-4a57-969f-c9e09d45f20e {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1213.487033] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7fe8a14c-66e6-42b7-88e3-7cfbd30644cf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.492433] env[62692]: DEBUG oslo_vmware.api [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142091, 'name': PowerOffVM_Task, 'duration_secs': 0.229624} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.493105] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1213.493982] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e08a237-5375-4f75-bd07-1ec9dee0de4a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.515126] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3053f8e3-12dd-4248-aa32-589eb6aaf4f6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.517750] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Created directory with path [datastore1] devstack-image-cache_base/c0e34614-043a-4a57-969f-c9e09d45f20e {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1213.517940] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_cad11812-0190-4a3b-a482-e1b2a143692e/OSTACK_IMG_cad11812-0190-4a3b-a482-e1b2a143692e.vmdk to [datastore1] devstack-image-cache_base/c0e34614-043a-4a57-969f-c9e09d45f20e/c0e34614-043a-4a57-969f-c9e09d45f20e.vmdk. {{(pid=62692) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1213.518196] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-6a7425a6-9cd1-4bef-ab3e-f59855fb0219 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.526635] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1213.526635] env[62692]: value = "task-1142093" [ 1213.526635] env[62692]: _type = "Task" [ 1213.526635] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.536711] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142093, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.570015] env[62692]: DEBUG nova.scheduler.client.report [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1213.821308] env[62692]: DEBUG nova.compute.manager [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1213.848067] env[62692]: DEBUG nova.virt.hardware [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1213.848360] env[62692]: DEBUG nova.virt.hardware [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1213.848524] env[62692]: DEBUG nova.virt.hardware [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1213.848700] env[62692]: DEBUG nova.virt.hardware [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1213.848894] env[62692]: DEBUG nova.virt.hardware [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1213.849068] env[62692]: DEBUG nova.virt.hardware [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1213.849313] env[62692]: DEBUG nova.virt.hardware [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1213.849484] env[62692]: DEBUG nova.virt.hardware [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1213.849663] env[62692]: DEBUG nova.virt.hardware [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1213.849839] env[62692]: DEBUG nova.virt.hardware [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1213.850441] env[62692]: DEBUG nova.virt.hardware [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1213.851465] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e400d5c2-938b-4b37-bf29-c383f636df66 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.861556] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6118ff8c-8d64-4e81-bc22-cda75ae7707f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.028565] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Creating Snapshot of the VM instance {{(pid=62692) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1214.028959] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8e44dfec-df7f-49dd-bd5f-ce1ed1833a8e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.040443] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142093, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.042042] env[62692]: DEBUG oslo_vmware.api [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1214.042042] env[62692]: value = "task-1142094" [ 1214.042042] env[62692]: _type = "Task" [ 1214.042042] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.050477] env[62692]: DEBUG oslo_vmware.api [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142094, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.075638] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.290s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1214.076334] env[62692]: DEBUG nova.compute.manager [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1214.078992] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a52630f8-0a43-4cf1-9a0b-538bd76b3b85 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 7.276s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1214.095389] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "e2383a6a-3581-40fc-a0eb-6981acdbf54a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1214.095692] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "e2383a6a-3581-40fc-a0eb-6981acdbf54a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1214.095936] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "e2383a6a-3581-40fc-a0eb-6981acdbf54a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1214.096149] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "e2383a6a-3581-40fc-a0eb-6981acdbf54a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1214.096331] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "e2383a6a-3581-40fc-a0eb-6981acdbf54a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1214.099301] env[62692]: INFO nova.compute.manager [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Terminating instance [ 1214.101262] env[62692]: DEBUG nova.compute.manager [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1214.101471] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1214.102328] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59493682-7415-4c45-81c8-93c83a49a46d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.110574] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1214.110901] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e2a26a95-c39f-4866-b741-652a5ff72a4a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.117264] env[62692]: DEBUG oslo_vmware.api [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1214.117264] env[62692]: value = "task-1142095" [ 1214.117264] env[62692]: _type = "Task" [ 1214.117264] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.127293] env[62692]: DEBUG oslo_vmware.api [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1142095, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.159804] env[62692]: DEBUG oslo_concurrency.lockutils [None req-917ddd32-8ccb-4465-85b9-50e7f944d2c6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "45eccffc-f059-4bef-97da-dacb16f20f88" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1214.160159] env[62692]: DEBUG oslo_concurrency.lockutils [None req-917ddd32-8ccb-4465-85b9-50e7f944d2c6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "45eccffc-f059-4bef-97da-dacb16f20f88" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1214.160442] env[62692]: INFO nova.compute.manager [None req-917ddd32-8ccb-4465-85b9-50e7f944d2c6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Attaching volume 8314cd71-4b96-4fcc-a78e-814d3eea76e0 to /dev/sdc [ 1214.199317] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb35fa8-7283-4889-8dfd-cf48d85a39be {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.206113] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5c284ce-c5f4-4920-922a-465b8b80024d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.220374] env[62692]: DEBUG nova.virt.block_device [None req-917ddd32-8ccb-4465-85b9-50e7f944d2c6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Updating existing volume attachment record: b7892ee1-14d4-4b3e-9860-cafe50691784 {{(pid=62692) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1214.287733] env[62692]: DEBUG nova.compute.manager [req-19563604-ae4d-463d-945d-d22e7f9863be req-ff2c4022-8606-4dd4-add0-8dc101e92282 service nova] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Received event network-vif-plugged-e7e22660-7340-4f0b-8d51-ad4b8834c59b {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1214.288042] env[62692]: DEBUG oslo_concurrency.lockutils [req-19563604-ae4d-463d-945d-d22e7f9863be req-ff2c4022-8606-4dd4-add0-8dc101e92282 service nova] Acquiring lock "29910848-22f2-4c2e-98fe-cd9e32e97865-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1214.288394] env[62692]: DEBUG oslo_concurrency.lockutils [req-19563604-ae4d-463d-945d-d22e7f9863be req-ff2c4022-8606-4dd4-add0-8dc101e92282 service nova] Lock "29910848-22f2-4c2e-98fe-cd9e32e97865-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1214.288394] env[62692]: DEBUG oslo_concurrency.lockutils [req-19563604-ae4d-463d-945d-d22e7f9863be req-ff2c4022-8606-4dd4-add0-8dc101e92282 service nova] Lock "29910848-22f2-4c2e-98fe-cd9e32e97865-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1214.288605] env[62692]: DEBUG nova.compute.manager [req-19563604-ae4d-463d-945d-d22e7f9863be req-ff2c4022-8606-4dd4-add0-8dc101e92282 service nova] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] No waiting events found dispatching network-vif-plugged-e7e22660-7340-4f0b-8d51-ad4b8834c59b {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1214.288683] env[62692]: WARNING nova.compute.manager [req-19563604-ae4d-463d-945d-d22e7f9863be req-ff2c4022-8606-4dd4-add0-8dc101e92282 service nova] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Received unexpected event network-vif-plugged-e7e22660-7340-4f0b-8d51-ad4b8834c59b for instance with vm_state building and task_state spawning. [ 1214.345741] env[62692]: DEBUG nova.network.neutron [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Successfully updated port: e7e22660-7340-4f0b-8d51-ad4b8834c59b {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1214.543391] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142093, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.554615] env[62692]: DEBUG oslo_vmware.api [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142094, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.582993] env[62692]: DEBUG nova.compute.utils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1214.587071] env[62692]: DEBUG nova.compute.manager [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1214.587259] env[62692]: DEBUG nova.network.neutron [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1214.632093] env[62692]: DEBUG oslo_vmware.api [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1142095, 'name': PowerOffVM_Task, 'duration_secs': 0.215347} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.632384] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1214.632558] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1214.632855] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2d091b43-eac9-4763-b6b4-d736bfaa2b55 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.670191] env[62692]: DEBUG nova.policy [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b7d10dd39f064a23ae4fa8a8ec777d92', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5a1ab0259e314d22b8f36ba6a9f4d693', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1214.696325] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1214.696578] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1214.696770] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Deleting the datastore file [datastore2] e2383a6a-3581-40fc-a0eb-6981acdbf54a {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1214.697092] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-09104b08-65be-4c91-99d4-f410f7d7bf82 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.706787] env[62692]: DEBUG oslo_vmware.api [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for the task: (returnval){ [ 1214.706787] env[62692]: value = "task-1142098" [ 1214.706787] env[62692]: _type = "Task" [ 1214.706787] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.715908] env[62692]: DEBUG oslo_vmware.api [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1142098, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.763354] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb49a0c8-41a1-4287-a6a8-77b1f34dd788 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.773991] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d3666cb-f3cc-484d-a8ee-17f1b57133f1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.812259] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c43d463-e488-4e25-baaa-9553530befcb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.821311] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad8eda12-d369-4f7a-84b9-538b1b914b8f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.837919] env[62692]: DEBUG nova.compute.provider_tree [None req-a52630f8-0a43-4cf1-9a0b-538bd76b3b85 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1214.849184] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "refresh_cache-29910848-22f2-4c2e-98fe-cd9e32e97865" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1214.849502] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquired lock "refresh_cache-29910848-22f2-4c2e-98fe-cd9e32e97865" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1214.849502] env[62692]: DEBUG nova.network.neutron [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1215.043119] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142093, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.052164] env[62692]: DEBUG oslo_vmware.api [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142094, 'name': CreateSnapshot_Task, 'duration_secs': 0.927629} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.052427] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Created Snapshot of the VM instance {{(pid=62692) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1215.053244] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2fa3efc-9805-4f10-8a6a-ff81b4d46a39 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.088126] env[62692]: DEBUG nova.compute.manager [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1215.105841] env[62692]: DEBUG nova.network.neutron [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Successfully created port: 7f32f75c-64a7-4073-9dde-0a4510530607 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1215.217276] env[62692]: DEBUG oslo_vmware.api [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Task: {'id': task-1142098, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.271115} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.217565] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1215.217758] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1215.217941] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1215.218139] env[62692]: INFO nova.compute.manager [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1215.218433] env[62692]: DEBUG oslo.service.loopingcall [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1215.218638] env[62692]: DEBUG nova.compute.manager [-] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1215.218735] env[62692]: DEBUG nova.network.neutron [-] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1215.344843] env[62692]: DEBUG nova.scheduler.client.report [None req-a52630f8-0a43-4cf1-9a0b-538bd76b3b85 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1215.540839] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142093, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.572068] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Creating linked-clone VM from snapshot {{(pid=62692) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1215.574408] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-92af458a-5564-417a-aa53-c9a29339a0c1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.576885] env[62692]: DEBUG nova.network.neutron [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1215.584800] env[62692]: DEBUG oslo_vmware.api [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1215.584800] env[62692]: value = "task-1142099" [ 1215.584800] env[62692]: _type = "Task" [ 1215.584800] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.599070] env[62692]: DEBUG oslo_vmware.api [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142099, 'name': CloneVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.740028] env[62692]: DEBUG nova.network.neutron [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Updating instance_info_cache with network_info: [{"id": "e7e22660-7340-4f0b-8d51-ad4b8834c59b", "address": "fa:16:3e:72:0b:83", "network": {"id": "8d7f8236-a3f7-4ce3-a9b4-c90747b73b30", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-289766533-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a1ab0259e314d22b8f36ba6a9f4d693", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7e22660-73", "ovs_interfaceid": "e7e22660-7340-4f0b-8d51-ad4b8834c59b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1216.040945] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142093, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.396019} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.041304] env[62692]: INFO nova.virt.vmwareapi.ds_util [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_cad11812-0190-4a3b-a482-e1b2a143692e/OSTACK_IMG_cad11812-0190-4a3b-a482-e1b2a143692e.vmdk to [datastore1] devstack-image-cache_base/c0e34614-043a-4a57-969f-c9e09d45f20e/c0e34614-043a-4a57-969f-c9e09d45f20e.vmdk. [ 1216.041498] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Cleaning up location [datastore1] OSTACK_IMG_cad11812-0190-4a3b-a482-e1b2a143692e {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1216.041673] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_cad11812-0190-4a3b-a482-e1b2a143692e {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1216.041924] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bfb0358f-7439-4146-992e-5f2996ad2784 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.048085] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1216.048085] env[62692]: value = "task-1142100" [ 1216.048085] env[62692]: _type = "Task" [ 1216.048085] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.055696] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142100, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.094860] env[62692]: DEBUG oslo_vmware.api [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142099, 'name': CloneVM_Task} progress is 94%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.101124] env[62692]: DEBUG nova.compute.manager [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1216.131843] env[62692]: DEBUG nova.virt.hardware [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1216.132283] env[62692]: DEBUG nova.virt.hardware [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1216.132486] env[62692]: DEBUG nova.virt.hardware [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1216.132809] env[62692]: DEBUG nova.virt.hardware [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1216.133039] env[62692]: DEBUG nova.virt.hardware [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1216.133249] env[62692]: DEBUG nova.virt.hardware [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1216.133497] env[62692]: DEBUG nova.virt.hardware [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1216.133693] env[62692]: DEBUG nova.virt.hardware [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1216.133892] env[62692]: DEBUG nova.virt.hardware [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1216.134107] env[62692]: DEBUG nova.virt.hardware [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1216.134370] env[62692]: DEBUG nova.virt.hardware [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1216.134776] env[62692]: DEBUG nova.network.neutron [-] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1216.136477] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05fb992c-25ec-4b3e-942c-40e0fe61e0d4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.145104] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d6331cf-7626-4f2b-b167-a35137a2e322 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.243311] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Releasing lock "refresh_cache-29910848-22f2-4c2e-98fe-cd9e32e97865" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1216.243639] env[62692]: DEBUG nova.compute.manager [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Instance network_info: |[{"id": "e7e22660-7340-4f0b-8d51-ad4b8834c59b", "address": "fa:16:3e:72:0b:83", "network": {"id": "8d7f8236-a3f7-4ce3-a9b4-c90747b73b30", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-289766533-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a1ab0259e314d22b8f36ba6a9f4d693", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7e22660-73", "ovs_interfaceid": "e7e22660-7340-4f0b-8d51-ad4b8834c59b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1216.244122] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:72:0b:83', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c1520c99-af74-4d61-a8ae-56aef56ef4f0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e7e22660-7340-4f0b-8d51-ad4b8834c59b', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1216.251594] env[62692]: DEBUG oslo.service.loopingcall [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1216.251807] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1216.252051] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bc6943c7-70d0-4ec4-9629-ffee78be8cf0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.272895] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1216.272895] env[62692]: value = "task-1142101" [ 1216.272895] env[62692]: _type = "Task" [ 1216.272895] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.279816] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1142101, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.318220] env[62692]: DEBUG nova.compute.manager [req-886b462e-d88b-4c64-8b13-ca13250bd31e req-867b9371-cf6f-4126-9d00-7362d111ae1e service nova] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Received event network-changed-e7e22660-7340-4f0b-8d51-ad4b8834c59b {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1216.318421] env[62692]: DEBUG nova.compute.manager [req-886b462e-d88b-4c64-8b13-ca13250bd31e req-867b9371-cf6f-4126-9d00-7362d111ae1e service nova] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Refreshing instance network info cache due to event network-changed-e7e22660-7340-4f0b-8d51-ad4b8834c59b. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1216.318701] env[62692]: DEBUG oslo_concurrency.lockutils [req-886b462e-d88b-4c64-8b13-ca13250bd31e req-867b9371-cf6f-4126-9d00-7362d111ae1e service nova] Acquiring lock "refresh_cache-29910848-22f2-4c2e-98fe-cd9e32e97865" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1216.318881] env[62692]: DEBUG oslo_concurrency.lockutils [req-886b462e-d88b-4c64-8b13-ca13250bd31e req-867b9371-cf6f-4126-9d00-7362d111ae1e service nova] Acquired lock "refresh_cache-29910848-22f2-4c2e-98fe-cd9e32e97865" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1216.319441] env[62692]: DEBUG nova.network.neutron [req-886b462e-d88b-4c64-8b13-ca13250bd31e req-867b9371-cf6f-4126-9d00-7362d111ae1e service nova] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Refreshing network info cache for port e7e22660-7340-4f0b-8d51-ad4b8834c59b {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1216.357032] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a52630f8-0a43-4cf1-9a0b-538bd76b3b85 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.278s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1216.357210] env[62692]: DEBUG nova.compute.manager [None req-a52630f8-0a43-4cf1-9a0b-538bd76b3b85 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=62692) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4909}} [ 1216.359897] env[62692]: DEBUG oslo_concurrency.lockutils [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.403s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1216.360143] env[62692]: DEBUG nova.objects.instance [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lazy-loading 'resources' on Instance uuid 0b7ea87c-252a-48b3-a1cd-c81e91362812 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1216.559293] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142100, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.043956} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.559780] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1216.560146] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0e34614-043a-4a57-969f-c9e09d45f20e/c0e34614-043a-4a57-969f-c9e09d45f20e.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1216.560607] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0e34614-043a-4a57-969f-c9e09d45f20e/c0e34614-043a-4a57-969f-c9e09d45f20e.vmdk to [datastore1] c2f2a485-363e-41af-9b3f-37092947ca08/c2f2a485-363e-41af-9b3f-37092947ca08.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1216.561017] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f8c2d793-76ad-4798-8519-6d93ec4fa9a0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.570255] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1216.570255] env[62692]: value = "task-1142103" [ 1216.570255] env[62692]: _type = "Task" [ 1216.570255] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.580344] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142103, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.596396] env[62692]: DEBUG oslo_vmware.api [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142099, 'name': CloneVM_Task} progress is 94%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.640170] env[62692]: INFO nova.compute.manager [-] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Took 1.42 seconds to deallocate network for instance. [ 1216.787060] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1142101, 'name': CreateVM_Task, 'duration_secs': 0.436994} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.787060] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1216.787060] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1216.787060] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1216.787060] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1216.787586] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c847abbf-954e-4e75-812d-7dd422f75a85 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.794071] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1216.794071] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5272126e-2a77-fd1e-d3e9-ab6a02a6d56e" [ 1216.794071] env[62692]: _type = "Task" [ 1216.794071] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.805356] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5272126e-2a77-fd1e-d3e9-ab6a02a6d56e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.917262] env[62692]: DEBUG nova.network.neutron [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Successfully updated port: 7f32f75c-64a7-4073-9dde-0a4510530607 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1216.945509] env[62692]: INFO nova.scheduler.client.report [None req-a52630f8-0a43-4cf1-9a0b-538bd76b3b85 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Deleted allocation for migration 1ab606dc-82f8-4a56-88d8-27a18eba60fc [ 1217.051581] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f0f8311-abc3-4ba3-84ed-2b26d85c8632 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.062380] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34018289-5da3-4bc3-9c52-530466853ee2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.109405] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cc1def3-64ac-4c14-ae0d-f71261182a52 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.115425] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142103, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.121095] env[62692]: DEBUG oslo_vmware.api [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142099, 'name': CloneVM_Task} progress is 95%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.124655] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-982f5dfe-b443-4378-8a7e-0fdb245fd239 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.140704] env[62692]: DEBUG nova.compute.provider_tree [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1217.143611] env[62692]: DEBUG nova.network.neutron [req-886b462e-d88b-4c64-8b13-ca13250bd31e req-867b9371-cf6f-4126-9d00-7362d111ae1e service nova] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Updated VIF entry in instance network info cache for port e7e22660-7340-4f0b-8d51-ad4b8834c59b. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1217.143611] env[62692]: DEBUG nova.network.neutron [req-886b462e-d88b-4c64-8b13-ca13250bd31e req-867b9371-cf6f-4126-9d00-7362d111ae1e service nova] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Updating instance_info_cache with network_info: [{"id": "e7e22660-7340-4f0b-8d51-ad4b8834c59b", "address": "fa:16:3e:72:0b:83", "network": {"id": "8d7f8236-a3f7-4ce3-a9b4-c90747b73b30", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-289766533-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a1ab0259e314d22b8f36ba6a9f4d693", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7e22660-73", "ovs_interfaceid": "e7e22660-7340-4f0b-8d51-ad4b8834c59b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1217.150024] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1217.311209] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5272126e-2a77-fd1e-d3e9-ab6a02a6d56e, 'name': SearchDatastore_Task, 'duration_secs': 0.030429} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.311794] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1217.312182] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1217.312595] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1217.312933] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1217.313266] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1217.313712] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-022312ef-e6ad-46ea-a75b-399d3d1e8dc6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.325169] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1217.325169] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1217.325848] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bced4d09-ad98-4aa4-801d-c7f6ef98cb96 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.333399] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1217.333399] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52652e1a-f9cb-c48e-0940-e12609894111" [ 1217.333399] env[62692]: _type = "Task" [ 1217.333399] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.342751] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52652e1a-f9cb-c48e-0940-e12609894111, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.422941] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "refresh_cache-0806e8c3-8b51-45e8-aaf8-cbd6f03f5489" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1217.423301] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquired lock "refresh_cache-0806e8c3-8b51-45e8-aaf8-cbd6f03f5489" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1217.423618] env[62692]: DEBUG nova.network.neutron [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1217.451750] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a52630f8-0a43-4cf1-9a0b-538bd76b3b85 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "90932f05-fad3-4c6d-87ae-ab059351b0be" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 13.597s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1217.583259] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142103, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.615789] env[62692]: DEBUG oslo_vmware.api [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142099, 'name': CloneVM_Task, 'duration_secs': 1.702622} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.616117] env[62692]: INFO nova.virt.vmwareapi.vmops [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Created linked-clone VM from snapshot [ 1217.616898] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f4c04b6-87e4-439f-a085-2a0fde394275 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.626953] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Uploading image a9c559b8-09bf-41ff-8b1b-16f309dd363f {{(pid=62692) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1217.645906] env[62692]: DEBUG nova.scheduler.client.report [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1217.649911] env[62692]: DEBUG oslo_concurrency.lockutils [req-886b462e-d88b-4c64-8b13-ca13250bd31e req-867b9371-cf6f-4126-9d00-7362d111ae1e service nova] Releasing lock "refresh_cache-29910848-22f2-4c2e-98fe-cd9e32e97865" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1217.650249] env[62692]: DEBUG nova.compute.manager [req-886b462e-d88b-4c64-8b13-ca13250bd31e req-867b9371-cf6f-4126-9d00-7362d111ae1e service nova] [instance: e2383a6a-3581-40fc-a0eb-6981acdbf54a] Received event network-vif-deleted-a291ddc4-8996-4755-b347-066f38e6ec7d {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1217.659576] env[62692]: DEBUG oslo_vmware.rw_handles [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1217.659576] env[62692]: value = "vm-249096" [ 1217.659576] env[62692]: _type = "VirtualMachine" [ 1217.659576] env[62692]: }. {{(pid=62692) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1217.659883] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-71d0b197-fb02-4a71-9c29-b6b9da923a34 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.671050] env[62692]: DEBUG oslo_vmware.rw_handles [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lease: (returnval){ [ 1217.671050] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]529593cf-bc9f-eaae-4a88-c2869e2b8073" [ 1217.671050] env[62692]: _type = "HttpNfcLease" [ 1217.671050] env[62692]: } obtained for exporting VM: (result){ [ 1217.671050] env[62692]: value = "vm-249096" [ 1217.671050] env[62692]: _type = "VirtualMachine" [ 1217.671050] env[62692]: }. {{(pid=62692) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1217.671396] env[62692]: DEBUG oslo_vmware.api [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the lease: (returnval){ [ 1217.671396] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]529593cf-bc9f-eaae-4a88-c2869e2b8073" [ 1217.671396] env[62692]: _type = "HttpNfcLease" [ 1217.671396] env[62692]: } to be ready. {{(pid=62692) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1217.680616] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1217.680616] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]529593cf-bc9f-eaae-4a88-c2869e2b8073" [ 1217.680616] env[62692]: _type = "HttpNfcLease" [ 1217.680616] env[62692]: } is initializing. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1217.846499] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52652e1a-f9cb-c48e-0940-e12609894111, 'name': SearchDatastore_Task, 'duration_secs': 0.013227} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.847406] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e915fc7-3c47-4e3d-903f-0130ad661e2b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.854745] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1217.854745] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5230eadf-f15e-1cfa-fcc9-e571fc57ffcc" [ 1217.854745] env[62692]: _type = "Task" [ 1217.854745] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.865806] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5230eadf-f15e-1cfa-fcc9-e571fc57ffcc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.885084] env[62692]: DEBUG nova.objects.instance [None req-56a181aa-e919-4e82-9ef2-e66f81b4ee78 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lazy-loading 'flavor' on Instance uuid 90932f05-fad3-4c6d-87ae-ab059351b0be {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1217.961958] env[62692]: DEBUG nova.network.neutron [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1218.086520] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142103, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.124222] env[62692]: DEBUG nova.network.neutron [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Updating instance_info_cache with network_info: [{"id": "7f32f75c-64a7-4073-9dde-0a4510530607", "address": "fa:16:3e:2a:0a:b8", "network": {"id": "8d7f8236-a3f7-4ce3-a9b4-c90747b73b30", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-289766533-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a1ab0259e314d22b8f36ba6a9f4d693", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f32f75c-64", "ovs_interfaceid": "7f32f75c-64a7-4073-9dde-0a4510530607", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1218.153965] env[62692]: DEBUG oslo_concurrency.lockutils [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.794s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1218.156460] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.008s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1218.156717] env[62692]: DEBUG nova.objects.instance [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lazy-loading 'resources' on Instance uuid e2383a6a-3581-40fc-a0eb-6981acdbf54a {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1218.178153] env[62692]: INFO nova.scheduler.client.report [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Deleted allocations for instance 0b7ea87c-252a-48b3-a1cd-c81e91362812 [ 1218.183859] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1218.183859] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]529593cf-bc9f-eaae-4a88-c2869e2b8073" [ 1218.183859] env[62692]: _type = "HttpNfcLease" [ 1218.183859] env[62692]: } is ready. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1218.184207] env[62692]: DEBUG oslo_vmware.rw_handles [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1218.184207] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]529593cf-bc9f-eaae-4a88-c2869e2b8073" [ 1218.184207] env[62692]: _type = "HttpNfcLease" [ 1218.184207] env[62692]: }. {{(pid=62692) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1218.185059] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-672f49c4-08b8-4dcd-a7da-fec9e5487008 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.194310] env[62692]: DEBUG oslo_vmware.rw_handles [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5222e1d7-9686-159c-8e2b-9767be4f09dc/disk-0.vmdk from lease info. {{(pid=62692) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1218.194523] env[62692]: DEBUG oslo_vmware.rw_handles [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5222e1d7-9686-159c-8e2b-9767be4f09dc/disk-0.vmdk for reading. {{(pid=62692) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1218.294371] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-4f1bb7be-c174-4765-a52e-9676edfe9bf8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.347548] env[62692]: DEBUG nova.compute.manager [req-f5951925-fd48-419c-8cce-51fe1ca1c1f0 req-c3c7aa89-921f-4122-a0da-92e19f290460 service nova] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Received event network-vif-plugged-7f32f75c-64a7-4073-9dde-0a4510530607 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1218.347889] env[62692]: DEBUG oslo_concurrency.lockutils [req-f5951925-fd48-419c-8cce-51fe1ca1c1f0 req-c3c7aa89-921f-4122-a0da-92e19f290460 service nova] Acquiring lock "0806e8c3-8b51-45e8-aaf8-cbd6f03f5489-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1218.348021] env[62692]: DEBUG oslo_concurrency.lockutils [req-f5951925-fd48-419c-8cce-51fe1ca1c1f0 req-c3c7aa89-921f-4122-a0da-92e19f290460 service nova] Lock "0806e8c3-8b51-45e8-aaf8-cbd6f03f5489-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1218.348181] env[62692]: DEBUG oslo_concurrency.lockutils [req-f5951925-fd48-419c-8cce-51fe1ca1c1f0 req-c3c7aa89-921f-4122-a0da-92e19f290460 service nova] Lock "0806e8c3-8b51-45e8-aaf8-cbd6f03f5489-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1218.348420] env[62692]: DEBUG nova.compute.manager [req-f5951925-fd48-419c-8cce-51fe1ca1c1f0 req-c3c7aa89-921f-4122-a0da-92e19f290460 service nova] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] No waiting events found dispatching network-vif-plugged-7f32f75c-64a7-4073-9dde-0a4510530607 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1218.348514] env[62692]: WARNING nova.compute.manager [req-f5951925-fd48-419c-8cce-51fe1ca1c1f0 req-c3c7aa89-921f-4122-a0da-92e19f290460 service nova] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Received unexpected event network-vif-plugged-7f32f75c-64a7-4073-9dde-0a4510530607 for instance with vm_state building and task_state spawning. [ 1218.348674] env[62692]: DEBUG nova.compute.manager [req-f5951925-fd48-419c-8cce-51fe1ca1c1f0 req-c3c7aa89-921f-4122-a0da-92e19f290460 service nova] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Received event network-changed-7f32f75c-64a7-4073-9dde-0a4510530607 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1218.348828] env[62692]: DEBUG nova.compute.manager [req-f5951925-fd48-419c-8cce-51fe1ca1c1f0 req-c3c7aa89-921f-4122-a0da-92e19f290460 service nova] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Refreshing instance network info cache due to event network-changed-7f32f75c-64a7-4073-9dde-0a4510530607. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1218.348993] env[62692]: DEBUG oslo_concurrency.lockutils [req-f5951925-fd48-419c-8cce-51fe1ca1c1f0 req-c3c7aa89-921f-4122-a0da-92e19f290460 service nova] Acquiring lock "refresh_cache-0806e8c3-8b51-45e8-aaf8-cbd6f03f5489" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1218.366732] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5230eadf-f15e-1cfa-fcc9-e571fc57ffcc, 'name': SearchDatastore_Task, 'duration_secs': 0.0149} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1218.367386] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1218.367665] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 29910848-22f2-4c2e-98fe-cd9e32e97865/29910848-22f2-4c2e-98fe-cd9e32e97865.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1218.367996] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-92957158-b394-44b2-b4ef-81b72c696d0d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.379982] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1218.379982] env[62692]: value = "task-1142105" [ 1218.379982] env[62692]: _type = "Task" [ 1218.379982] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.393651] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142105, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.394226] env[62692]: DEBUG oslo_concurrency.lockutils [None req-56a181aa-e919-4e82-9ef2-e66f81b4ee78 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "refresh_cache-90932f05-fad3-4c6d-87ae-ab059351b0be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1218.394390] env[62692]: DEBUG oslo_concurrency.lockutils [None req-56a181aa-e919-4e82-9ef2-e66f81b4ee78 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquired lock "refresh_cache-90932f05-fad3-4c6d-87ae-ab059351b0be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1218.394565] env[62692]: DEBUG nova.network.neutron [None req-56a181aa-e919-4e82-9ef2-e66f81b4ee78 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1218.394749] env[62692]: DEBUG nova.objects.instance [None req-56a181aa-e919-4e82-9ef2-e66f81b4ee78 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lazy-loading 'info_cache' on Instance uuid 90932f05-fad3-4c6d-87ae-ab059351b0be {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1218.585700] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142103, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.627548] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Releasing lock "refresh_cache-0806e8c3-8b51-45e8-aaf8-cbd6f03f5489" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1218.627929] env[62692]: DEBUG nova.compute.manager [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Instance network_info: |[{"id": "7f32f75c-64a7-4073-9dde-0a4510530607", "address": "fa:16:3e:2a:0a:b8", "network": {"id": "8d7f8236-a3f7-4ce3-a9b4-c90747b73b30", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-289766533-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a1ab0259e314d22b8f36ba6a9f4d693", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f32f75c-64", "ovs_interfaceid": "7f32f75c-64a7-4073-9dde-0a4510530607", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1218.628474] env[62692]: DEBUG oslo_concurrency.lockutils [req-f5951925-fd48-419c-8cce-51fe1ca1c1f0 req-c3c7aa89-921f-4122-a0da-92e19f290460 service nova] Acquired lock "refresh_cache-0806e8c3-8b51-45e8-aaf8-cbd6f03f5489" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1218.628824] env[62692]: DEBUG nova.network.neutron [req-f5951925-fd48-419c-8cce-51fe1ca1c1f0 req-c3c7aa89-921f-4122-a0da-92e19f290460 service nova] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Refreshing network info cache for port 7f32f75c-64a7-4073-9dde-0a4510530607 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1218.630471] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2a:0a:b8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c1520c99-af74-4d61-a8ae-56aef56ef4f0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7f32f75c-64a7-4073-9dde-0a4510530607', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1218.639999] env[62692]: DEBUG oslo.service.loopingcall [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1218.641281] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1218.641547] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6e9a2e40-1a0f-42b2-abf1-c9667cf4c5fc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.671285] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1218.671285] env[62692]: value = "task-1142106" [ 1218.671285] env[62692]: _type = "Task" [ 1218.671285] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.679667] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1142106, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.685697] env[62692]: DEBUG oslo_concurrency.lockutils [None req-955f4baf-0dab-4148-b216-f202002fc67d tempest-AttachVolumeNegativeTest-786092174 tempest-AttachVolumeNegativeTest-786092174-project-member] Lock "0b7ea87c-252a-48b3-a1cd-c81e91362812" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.021s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1218.774040] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-917ddd32-8ccb-4465-85b9-50e7f944d2c6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Volume attach. Driver type: vmdk {{(pid=62692) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1218.774566] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-917ddd32-8ccb-4465-85b9-50e7f944d2c6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249094', 'volume_id': '8314cd71-4b96-4fcc-a78e-814d3eea76e0', 'name': 'volume-8314cd71-4b96-4fcc-a78e-814d3eea76e0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '45eccffc-f059-4bef-97da-dacb16f20f88', 'attached_at': '', 'detached_at': '', 'volume_id': '8314cd71-4b96-4fcc-a78e-814d3eea76e0', 'serial': '8314cd71-4b96-4fcc-a78e-814d3eea76e0'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1218.776091] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80387ac3-a80f-4d87-a38f-6131e43b5742 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.800697] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57f3fac9-b60f-4669-af78-c26b82526952 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.836551] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-917ddd32-8ccb-4465-85b9-50e7f944d2c6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] volume-8314cd71-4b96-4fcc-a78e-814d3eea76e0/volume-8314cd71-4b96-4fcc-a78e-814d3eea76e0.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1218.838377] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-16fa311b-b2ab-4e48-aa4b-134a180557ee {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.856258] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-251345ee-4654-4184-88f6-c0809ff11c52 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.869981] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-440031a9-70b0-4eeb-9508-e840498e7faf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.874855] env[62692]: DEBUG oslo_vmware.api [None req-917ddd32-8ccb-4465-85b9-50e7f944d2c6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1218.874855] env[62692]: value = "task-1142108" [ 1218.874855] env[62692]: _type = "Task" [ 1218.874855] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.915189] env[62692]: DEBUG nova.objects.base [None req-56a181aa-e919-4e82-9ef2-e66f81b4ee78 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Object Instance<90932f05-fad3-4c6d-87ae-ab059351b0be> lazy-loaded attributes: flavor,info_cache {{(pid=62692) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1218.922163] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4db72c16-6ad5-4b94-b633-7ea72b1019f7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.929248] env[62692]: DEBUG oslo_vmware.api [None req-917ddd32-8ccb-4465-85b9-50e7f944d2c6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1142108, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.936286] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142105, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.941108] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f36b8659-df44-4483-978b-f8a24556593e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.960044] env[62692]: DEBUG nova.compute.provider_tree [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1219.083975] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142103, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.38148} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.084387] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0e34614-043a-4a57-969f-c9e09d45f20e/c0e34614-043a-4a57-969f-c9e09d45f20e.vmdk to [datastore1] c2f2a485-363e-41af-9b3f-37092947ca08/c2f2a485-363e-41af-9b3f-37092947ca08.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1219.085444] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f616c3ba-6382-4770-8ea9-22271975a10c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.113087] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] c2f2a485-363e-41af-9b3f-37092947ca08/c2f2a485-363e-41af-9b3f-37092947ca08.vmdk or device None with type streamOptimized {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1219.113674] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b7edbbb-b76d-489c-913d-f393ae85e8eb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.134093] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1219.134093] env[62692]: value = "task-1142109" [ 1219.134093] env[62692]: _type = "Task" [ 1219.134093] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.144079] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142109, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.183782] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1142106, 'name': CreateVM_Task} progress is 25%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.389894] env[62692]: DEBUG oslo_vmware.api [None req-917ddd32-8ccb-4465-85b9-50e7f944d2c6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1142108, 'name': ReconfigVM_Task, 'duration_secs': 0.499497} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.390303] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-917ddd32-8ccb-4465-85b9-50e7f944d2c6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Reconfigured VM instance instance-0000006f to attach disk [datastore2] volume-8314cd71-4b96-4fcc-a78e-814d3eea76e0/volume-8314cd71-4b96-4fcc-a78e-814d3eea76e0.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1219.396480] env[62692]: DEBUG nova.network.neutron [req-f5951925-fd48-419c-8cce-51fe1ca1c1f0 req-c3c7aa89-921f-4122-a0da-92e19f290460 service nova] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Updated VIF entry in instance network info cache for port 7f32f75c-64a7-4073-9dde-0a4510530607. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1219.396840] env[62692]: DEBUG nova.network.neutron [req-f5951925-fd48-419c-8cce-51fe1ca1c1f0 req-c3c7aa89-921f-4122-a0da-92e19f290460 service nova] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Updating instance_info_cache with network_info: [{"id": "7f32f75c-64a7-4073-9dde-0a4510530607", "address": "fa:16:3e:2a:0a:b8", "network": {"id": "8d7f8236-a3f7-4ce3-a9b4-c90747b73b30", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-289766533-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a1ab0259e314d22b8f36ba6a9f4d693", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f32f75c-64", "ovs_interfaceid": "7f32f75c-64a7-4073-9dde-0a4510530607", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1219.399181] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b89a2767-07d1-4d85-8e7a-a37a114fa412 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.416173] env[62692]: DEBUG oslo_vmware.api [None req-917ddd32-8ccb-4465-85b9-50e7f944d2c6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1219.416173] env[62692]: value = "task-1142110" [ 1219.416173] env[62692]: _type = "Task" [ 1219.416173] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.430804] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142105, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.741285} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.434346] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 29910848-22f2-4c2e-98fe-cd9e32e97865/29910848-22f2-4c2e-98fe-cd9e32e97865.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1219.434602] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1219.434967] env[62692]: DEBUG oslo_vmware.api [None req-917ddd32-8ccb-4465-85b9-50e7f944d2c6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1142110, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.435902] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-de912372-ee95-4858-aa05-6f7a02713366 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.442025] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1219.442025] env[62692]: value = "task-1142111" [ 1219.442025] env[62692]: _type = "Task" [ 1219.442025] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.449755] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142111, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.462245] env[62692]: DEBUG nova.scheduler.client.report [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1219.644925] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142109, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.664679] env[62692]: DEBUG nova.network.neutron [None req-56a181aa-e919-4e82-9ef2-e66f81b4ee78 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Updating instance_info_cache with network_info: [{"id": "21e76034-6c0f-4938-afda-0d005cf45759", "address": "fa:16:3e:2d:c0:4a", "network": {"id": "77695f27-3027-4a92-8b9b-0a90cdc53f62", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-337681783-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697b4281872f4547ad3d051e644b95ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21e76034-6c", "ovs_interfaceid": "21e76034-6c0f-4938-afda-0d005cf45759", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1219.682909] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1142106, 'name': CreateVM_Task, 'duration_secs': 0.721139} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.683110] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1219.683840] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1219.684208] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1219.684776] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1219.685062] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f20ef60-6e66-4bce-8115-bfa1560337e1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.689771] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1219.689771] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52ccb86a-3926-c6b3-9a98-c205874c582e" [ 1219.689771] env[62692]: _type = "Task" [ 1219.689771] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.699817] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52ccb86a-3926-c6b3-9a98-c205874c582e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.911955] env[62692]: DEBUG oslo_concurrency.lockutils [req-f5951925-fd48-419c-8cce-51fe1ca1c1f0 req-c3c7aa89-921f-4122-a0da-92e19f290460 service nova] Releasing lock "refresh_cache-0806e8c3-8b51-45e8-aaf8-cbd6f03f5489" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1219.931283] env[62692]: DEBUG oslo_vmware.api [None req-917ddd32-8ccb-4465-85b9-50e7f944d2c6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1142110, 'name': ReconfigVM_Task, 'duration_secs': 0.149204} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.931665] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-917ddd32-8ccb-4465-85b9-50e7f944d2c6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249094', 'volume_id': '8314cd71-4b96-4fcc-a78e-814d3eea76e0', 'name': 'volume-8314cd71-4b96-4fcc-a78e-814d3eea76e0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '45eccffc-f059-4bef-97da-dacb16f20f88', 'attached_at': '', 'detached_at': '', 'volume_id': '8314cd71-4b96-4fcc-a78e-814d3eea76e0', 'serial': '8314cd71-4b96-4fcc-a78e-814d3eea76e0'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1219.951873] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142111, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.102721} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.952260] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1219.953154] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-738c4974-5b6a-463c-8354-28a883ff0da6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.967601] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.811s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1219.978333] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Reconfiguring VM instance instance-00000076 to attach disk [datastore2] 29910848-22f2-4c2e-98fe-cd9e32e97865/29910848-22f2-4c2e-98fe-cd9e32e97865.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1219.979504] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d05c9477-7044-4315-9515-2124f5b82c51 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.999404] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1219.999404] env[62692]: value = "task-1142112" [ 1219.999404] env[62692]: _type = "Task" [ 1219.999404] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.000600] env[62692]: INFO nova.scheduler.client.report [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Deleted allocations for instance e2383a6a-3581-40fc-a0eb-6981acdbf54a [ 1220.013398] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142112, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.145416] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142109, 'name': ReconfigVM_Task, 'duration_secs': 0.818704} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.145843] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Reconfigured VM instance instance-0000006a to attach disk [datastore1] c2f2a485-363e-41af-9b3f-37092947ca08/c2f2a485-363e-41af-9b3f-37092947ca08.vmdk or device None with type streamOptimized {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1220.146811] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'encryption_format': None, 'encryption_secret_uuid': None, 'guest_format': None, 'device_type': 'disk', 'device_name': '/dev/sda', 'size': 0, 'encrypted': False, 'encryption_options': None, 'disk_bus': None, 'boot_index': 0, 'image_id': '81eb64fb-5a0c-47b7-9948-7dcb01b911bd'}], 'ephemerals': [], 'block_device_mapping': [{'mount_device': '/dev/sdb', 'attachment_id': 'b92a48fc-b907-4b2f-9a4e-f8b9d6d96819', 'guest_format': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249089', 'volume_id': 'c77c2006-709b-42d0-a47c-106f2a3f7237', 'name': 'volume-c77c2006-709b-42d0-a47c-106f2a3f7237', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'c2f2a485-363e-41af-9b3f-37092947ca08', 'attached_at': '', 'detached_at': '', 'volume_id': 'c77c2006-709b-42d0-a47c-106f2a3f7237', 'serial': 'c77c2006-709b-42d0-a47c-106f2a3f7237'}, 'device_type': None, 'delete_on_termination': False, 'disk_bus': None, 'boot_index': None, 'volume_type': None}], 'swap': None} {{(pid=62692) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1220.147183] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Volume attach. Driver type: vmdk {{(pid=62692) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1220.148011] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249089', 'volume_id': 'c77c2006-709b-42d0-a47c-106f2a3f7237', 'name': 'volume-c77c2006-709b-42d0-a47c-106f2a3f7237', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'c2f2a485-363e-41af-9b3f-37092947ca08', 'attached_at': '', 'detached_at': '', 'volume_id': 'c77c2006-709b-42d0-a47c-106f2a3f7237', 'serial': 'c77c2006-709b-42d0-a47c-106f2a3f7237'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1220.148848] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2794b5d-33c5-4d28-b770-b9f5c88c74f5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.167062] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c14432d-3b37-4df0-97c2-e5ff7b1f2573 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.170066] env[62692]: DEBUG oslo_concurrency.lockutils [None req-56a181aa-e919-4e82-9ef2-e66f81b4ee78 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Releasing lock "refresh_cache-90932f05-fad3-4c6d-87ae-ab059351b0be" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1220.928864] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5790017f-8b93-4b97-abe4-e297f6ccba3f tempest-ServerActionsTestOtherA-1757498226 tempest-ServerActionsTestOtherA-1757498226-project-member] Lock "e2383a6a-3581-40fc-a0eb-6981acdbf54a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.833s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1220.930786] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-56a181aa-e919-4e82-9ef2-e66f81b4ee78 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1220.938379] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] volume-c77c2006-709b-42d0-a47c-106f2a3f7237/volume-c77c2006-709b-42d0-a47c-106f2a3f7237.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1220.945332] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d06b4c2a-34bc-475b-9d89-2bae92c8fb42 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.947226] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7d6f85e1-472f-4b30-a6c0-f8bd322500dd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.967405] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142112, 'name': ReconfigVM_Task, 'duration_secs': 0.265282} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.972508] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Reconfigured VM instance instance-00000076 to attach disk [datastore2] 29910848-22f2-4c2e-98fe-cd9e32e97865/29910848-22f2-4c2e-98fe-cd9e32e97865.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1220.973310] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1220.973310] env[62692]: value = "task-1142114" [ 1220.973310] env[62692]: _type = "Task" [ 1220.973310] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.973895] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52ccb86a-3926-c6b3-9a98-c205874c582e, 'name': SearchDatastore_Task, 'duration_secs': 0.01608} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.974218] env[62692]: DEBUG oslo_vmware.api [None req-56a181aa-e919-4e82-9ef2-e66f81b4ee78 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1220.974218] env[62692]: value = "task-1142113" [ 1220.974218] env[62692]: _type = "Task" [ 1220.974218] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.974462] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-27f9b7ef-8fb5-46e3-8bc7-a04c4fe1290b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.976332] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1220.976606] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1220.976897] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1220.977113] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1220.977334] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1220.980309] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-00ec58dc-e8b9-4702-8b18-ea15b825ef7a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.991877] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142114, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.998960] env[62692]: DEBUG oslo_vmware.api [None req-56a181aa-e919-4e82-9ef2-e66f81b4ee78 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142113, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.999533] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1220.999533] env[62692]: value = "task-1142115" [ 1220.999533] env[62692]: _type = "Task" [ 1220.999533] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.001204] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1221.001204] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1221.004800] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f47d2e34-340d-4f79-983c-96c336373387 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.015278] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142115, 'name': Rename_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.015963] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1221.015963] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]528cb5a6-5e30-b172-25bd-0f763e782073" [ 1221.015963] env[62692]: _type = "Task" [ 1221.015963] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.026983] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]528cb5a6-5e30-b172-25bd-0f763e782073, 'name': SearchDatastore_Task, 'duration_secs': 0.011315} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.027919] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4aa69d88-416d-4d48-86dc-42cba720681d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.033592] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1221.033592] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5259c186-bb18-65d6-6e8f-8b2f35a03cb3" [ 1221.033592] env[62692]: _type = "Task" [ 1221.033592] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.042463] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5259c186-bb18-65d6-6e8f-8b2f35a03cb3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.490022] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142114, 'name': ReconfigVM_Task, 'duration_secs': 0.384175} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.493254] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Reconfigured VM instance instance-0000006a to attach disk [datastore2] volume-c77c2006-709b-42d0-a47c-106f2a3f7237/volume-c77c2006-709b-42d0-a47c-106f2a3f7237.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1221.498590] env[62692]: DEBUG oslo_vmware.api [None req-56a181aa-e919-4e82-9ef2-e66f81b4ee78 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142113, 'name': PowerOnVM_Task, 'duration_secs': 0.390153} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.499508] env[62692]: DEBUG nova.objects.instance [None req-917ddd32-8ccb-4465-85b9-50e7f944d2c6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lazy-loading 'flavor' on Instance uuid 45eccffc-f059-4bef-97da-dacb16f20f88 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1221.501240] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ada519df-3598-4f84-a3b2-484d162a0fac {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.513382] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-56a181aa-e919-4e82-9ef2-e66f81b4ee78 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1221.513382] env[62692]: DEBUG nova.compute.manager [None req-56a181aa-e919-4e82-9ef2-e66f81b4ee78 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1221.514647] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d48c77a-1d2e-4f93-93f3-603f4d09aac1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.532706] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1221.532706] env[62692]: value = "task-1142116" [ 1221.532706] env[62692]: _type = "Task" [ 1221.532706] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.533351] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142115, 'name': Rename_Task, 'duration_secs': 0.13993} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.533562] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1221.538772] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eb33a668-e262-4e47-8101-bab89931169b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.549130] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142116, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.556029] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5259c186-bb18-65d6-6e8f-8b2f35a03cb3, 'name': SearchDatastore_Task, 'duration_secs': 0.011028} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.556029] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1221.556029] env[62692]: value = "task-1142117" [ 1221.556029] env[62692]: _type = "Task" [ 1221.556029] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.556029] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1221.556029] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489/0806e8c3-8b51-45e8-aaf8-cbd6f03f5489.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1221.556029] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e1df42dc-2f5f-4225-a997-f910f2effd6d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.567706] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142117, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.569301] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1221.569301] env[62692]: value = "task-1142118" [ 1221.569301] env[62692]: _type = "Task" [ 1221.569301] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.577678] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142118, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.018152] env[62692]: DEBUG oslo_concurrency.lockutils [None req-917ddd32-8ccb-4465-85b9-50e7f944d2c6 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "45eccffc-f059-4bef-97da-dacb16f20f88" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.858s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1222.054537] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142116, 'name': ReconfigVM_Task, 'duration_secs': 0.176072} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.055043] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249089', 'volume_id': 'c77c2006-709b-42d0-a47c-106f2a3f7237', 'name': 'volume-c77c2006-709b-42d0-a47c-106f2a3f7237', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'c2f2a485-363e-41af-9b3f-37092947ca08', 'attached_at': '', 'detached_at': '', 'volume_id': 'c77c2006-709b-42d0-a47c-106f2a3f7237', 'serial': 'c77c2006-709b-42d0-a47c-106f2a3f7237'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1222.056381] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a58d206f-caa4-42eb-a52c-64c49bfe6608 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.081419] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142117, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.081419] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1222.081419] env[62692]: value = "task-1142120" [ 1222.081419] env[62692]: _type = "Task" [ 1222.081419] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.089560] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142118, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.093137] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142120, 'name': Rename_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.574027] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142117, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.582841] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142118, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.575762} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.589480] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489/0806e8c3-8b51-45e8-aaf8-cbd6f03f5489.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1222.589480] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1222.589480] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cb75d3e1-f382-4cb5-918e-a1327c28261c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.595069] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142120, 'name': Rename_Task, 'duration_secs': 0.248967} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.599052] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1222.599052] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1222.599052] env[62692]: value = "task-1142121" [ 1222.599052] env[62692]: _type = "Task" [ 1222.599052] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.599052] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-346b66d4-d34b-4ec4-97cf-ace1346d8447 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.611266] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142121, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.615029] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1222.615029] env[62692]: value = "task-1142122" [ 1222.615029] env[62692]: _type = "Task" [ 1222.615029] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.620925] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142122, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.655678] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "90932f05-fad3-4c6d-87ae-ab059351b0be" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1222.655678] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "90932f05-fad3-4c6d-87ae-ab059351b0be" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1222.655678] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "90932f05-fad3-4c6d-87ae-ab059351b0be-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1222.655678] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "90932f05-fad3-4c6d-87ae-ab059351b0be-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1222.655678] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "90932f05-fad3-4c6d-87ae-ab059351b0be-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1222.658826] env[62692]: INFO nova.compute.manager [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Terminating instance [ 1222.662078] env[62692]: DEBUG nova.compute.manager [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1222.662078] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1222.663149] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-778640e6-3a82-4789-90e8-0c06c5651be2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.671274] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1222.672033] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e51dbcc7-872d-46eb-aa56-c3ba1a73d24a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.678106] env[62692]: DEBUG oslo_vmware.api [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1222.678106] env[62692]: value = "task-1142123" [ 1222.678106] env[62692]: _type = "Task" [ 1222.678106] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.687390] env[62692]: DEBUG oslo_vmware.api [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142123, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.841686] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f586c60d-e4f7-4fc4-bf04-f6b47b67da85 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "9d3c0909-6207-46b7-b83b-de4d7d626fd7" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1222.842045] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f586c60d-e4f7-4fc4-bf04-f6b47b67da85 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "9d3c0909-6207-46b7-b83b-de4d7d626fd7" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1222.842192] env[62692]: DEBUG nova.compute.manager [None req-f586c60d-e4f7-4fc4-bf04-f6b47b67da85 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1222.843237] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eb0d191-c479-4547-87ee-0109ce707bb5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.850929] env[62692]: DEBUG nova.compute.manager [None req-f586c60d-e4f7-4fc4-bf04-f6b47b67da85 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62692) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1222.851660] env[62692]: DEBUG nova.objects.instance [None req-f586c60d-e4f7-4fc4-bf04-f6b47b67da85 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lazy-loading 'flavor' on Instance uuid 9d3c0909-6207-46b7-b83b-de4d7d626fd7 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1222.960579] env[62692]: DEBUG oslo_concurrency.lockutils [None req-00ff7331-4269-47a7-8091-f048ed6ca116 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "45eccffc-f059-4bef-97da-dacb16f20f88" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1222.960913] env[62692]: DEBUG oslo_concurrency.lockutils [None req-00ff7331-4269-47a7-8091-f048ed6ca116 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "45eccffc-f059-4bef-97da-dacb16f20f88" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1223.071634] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142117, 'name': PowerOnVM_Task, 'duration_secs': 1.092739} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.071925] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1223.072180] env[62692]: INFO nova.compute.manager [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Took 9.25 seconds to spawn the instance on the hypervisor. [ 1223.072367] env[62692]: DEBUG nova.compute.manager [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1223.073241] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0985427c-065e-4b3c-970f-4bf1dab31cc3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.111573] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142121, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.136943} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.111925] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1223.112648] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2ac11e2-a714-4955-8008-3236c90a08a9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.123238] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142122, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.140645] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Reconfiguring VM instance instance-00000077 to attach disk [datastore2] 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489/0806e8c3-8b51-45e8-aaf8-cbd6f03f5489.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1223.141664] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b84a3088-75ad-4883-80ca-2067a436ba43 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.161412] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1223.161412] env[62692]: value = "task-1142124" [ 1223.161412] env[62692]: _type = "Task" [ 1223.161412] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.169545] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142124, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.189288] env[62692]: DEBUG oslo_vmware.api [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142123, 'name': PowerOffVM_Task, 'duration_secs': 0.191121} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.189638] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1223.189854] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1223.190451] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-376c84ba-dc43-4afa-81d5-9bb7ba6a8521 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.254745] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1223.254993] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1223.255256] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Deleting the datastore file [datastore2] 90932f05-fad3-4c6d-87ae-ab059351b0be {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1223.255535] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bca27fff-2a4b-4f8d-8135-ccf2f0411b9a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.262420] env[62692]: DEBUG oslo_vmware.api [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1223.262420] env[62692]: value = "task-1142126" [ 1223.262420] env[62692]: _type = "Task" [ 1223.262420] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.270448] env[62692]: DEBUG oslo_vmware.api [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142126, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.357562] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f586c60d-e4f7-4fc4-bf04-f6b47b67da85 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1223.357895] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e7c2620b-b321-4619-a9dc-d0dd0681e67a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.364589] env[62692]: DEBUG oslo_vmware.api [None req-f586c60d-e4f7-4fc4-bf04-f6b47b67da85 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1223.364589] env[62692]: value = "task-1142127" [ 1223.364589] env[62692]: _type = "Task" [ 1223.364589] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.372721] env[62692]: DEBUG oslo_vmware.api [None req-f586c60d-e4f7-4fc4-bf04-f6b47b67da85 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142127, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.464662] env[62692]: INFO nova.compute.manager [None req-00ff7331-4269-47a7-8091-f048ed6ca116 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Detaching volume 89113b2e-696c-4171-8d9f-479a6c1f41db [ 1223.503156] env[62692]: INFO nova.virt.block_device [None req-00ff7331-4269-47a7-8091-f048ed6ca116 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Attempting to driver detach volume 89113b2e-696c-4171-8d9f-479a6c1f41db from mountpoint /dev/sdb [ 1223.503452] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-00ff7331-4269-47a7-8091-f048ed6ca116 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Volume detach. Driver type: vmdk {{(pid=62692) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1223.503767] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-00ff7331-4269-47a7-8091-f048ed6ca116 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249091', 'volume_id': '89113b2e-696c-4171-8d9f-479a6c1f41db', 'name': 'volume-89113b2e-696c-4171-8d9f-479a6c1f41db', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '45eccffc-f059-4bef-97da-dacb16f20f88', 'attached_at': '', 'detached_at': '', 'volume_id': '89113b2e-696c-4171-8d9f-479a6c1f41db', 'serial': '89113b2e-696c-4171-8d9f-479a6c1f41db'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1223.504926] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99739559-de52-496e-9a64-bbe59d296f5d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.529965] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-186edf0b-e436-4fa0-859b-dba7ffa42c3e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.537744] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58af6d79-3e3c-4c3d-b602-dbedbdc2f93b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.560922] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34106b2b-bbb5-4d81-a7ea-b141267e0a0e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.577409] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-00ff7331-4269-47a7-8091-f048ed6ca116 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] The volume has not been displaced from its original location: [datastore2] volume-89113b2e-696c-4171-8d9f-479a6c1f41db/volume-89113b2e-696c-4171-8d9f-479a6c1f41db.vmdk. No consolidation needed. {{(pid=62692) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1223.583173] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-00ff7331-4269-47a7-8091-f048ed6ca116 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Reconfiguring VM instance instance-0000006f to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1223.587912] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5fc27c1e-5579-4ae4-81e1-1c0f05db36e2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.607315] env[62692]: INFO nova.compute.manager [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Took 18.72 seconds to build instance. [ 1223.613633] env[62692]: DEBUG oslo_vmware.api [None req-00ff7331-4269-47a7-8091-f048ed6ca116 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1223.613633] env[62692]: value = "task-1142128" [ 1223.613633] env[62692]: _type = "Task" [ 1223.613633] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.630017] env[62692]: DEBUG oslo_vmware.api [None req-00ff7331-4269-47a7-8091-f048ed6ca116 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1142128, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.633953] env[62692]: DEBUG oslo_vmware.api [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142122, 'name': PowerOnVM_Task, 'duration_secs': 0.635665} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.635130] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1223.674098] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142124, 'name': ReconfigVM_Task, 'duration_secs': 0.313414} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.674098] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Reconfigured VM instance instance-00000077 to attach disk [datastore2] 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489/0806e8c3-8b51-45e8-aaf8-cbd6f03f5489.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1223.674395] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-422aa470-9740-4189-8268-9ebacff51a0c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.681814] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1223.681814] env[62692]: value = "task-1142129" [ 1223.681814] env[62692]: _type = "Task" [ 1223.681814] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.693340] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142129, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.748188] env[62692]: DEBUG nova.compute.manager [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1223.749336] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4f01ece-ac97-46ab-a4c6-d07bf0acfcff {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.771830] env[62692]: DEBUG oslo_vmware.api [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142126, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.209412} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.772067] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1223.772269] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1223.772448] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1223.772628] env[62692]: INFO nova.compute.manager [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1223.772950] env[62692]: DEBUG oslo.service.loopingcall [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1223.773171] env[62692]: DEBUG nova.compute.manager [-] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1223.773263] env[62692]: DEBUG nova.network.neutron [-] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1223.873913] env[62692]: DEBUG oslo_vmware.api [None req-f586c60d-e4f7-4fc4-bf04-f6b47b67da85 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142127, 'name': PowerOffVM_Task, 'duration_secs': 0.332774} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.874220] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-f586c60d-e4f7-4fc4-bf04-f6b47b67da85 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1223.874538] env[62692]: DEBUG nova.compute.manager [None req-f586c60d-e4f7-4fc4-bf04-f6b47b67da85 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1223.875159] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e7c39f5-8cb6-4eb2-a5a9-349230cc748a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.110220] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "29910848-22f2-4c2e-98fe-cd9e32e97865" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.240s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1224.123954] env[62692]: DEBUG oslo_vmware.api [None req-00ff7331-4269-47a7-8091-f048ed6ca116 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1142128, 'name': ReconfigVM_Task, 'duration_secs': 0.356134} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.124288] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-00ff7331-4269-47a7-8091-f048ed6ca116 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Reconfigured VM instance instance-0000006f to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1224.129022] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da477f57-1d40-4cc2-b626-a060c42c7323 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.145165] env[62692]: DEBUG oslo_vmware.api [None req-00ff7331-4269-47a7-8091-f048ed6ca116 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1224.145165] env[62692]: value = "task-1142130" [ 1224.145165] env[62692]: _type = "Task" [ 1224.145165] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.153102] env[62692]: DEBUG oslo_vmware.api [None req-00ff7331-4269-47a7-8091-f048ed6ca116 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1142130, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.194477] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142129, 'name': Rename_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.268511] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29f97919-0424-44c4-98d8-90a0b4d84ce2 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "c2f2a485-363e-41af-9b3f-37092947ca08" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 29.070s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1224.386071] env[62692]: DEBUG oslo_concurrency.lockutils [None req-f586c60d-e4f7-4fc4-bf04-f6b47b67da85 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "9d3c0909-6207-46b7-b83b-de4d7d626fd7" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.544s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1224.424868] env[62692]: DEBUG nova.compute.manager [req-9fa71750-13e9-4716-ab16-20193e4594f3 req-b3c2f81e-3715-4ecf-b2fd-b1659bd5118d service nova] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Received event network-vif-deleted-21e76034-6c0f-4938-afda-0d005cf45759 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1224.424972] env[62692]: INFO nova.compute.manager [req-9fa71750-13e9-4716-ab16-20193e4594f3 req-b3c2f81e-3715-4ecf-b2fd-b1659bd5118d service nova] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Neutron deleted interface 21e76034-6c0f-4938-afda-0d005cf45759; detaching it from the instance and deleting it from the info cache [ 1224.425213] env[62692]: DEBUG nova.network.neutron [req-9fa71750-13e9-4716-ab16-20193e4594f3 req-b3c2f81e-3715-4ecf-b2fd-b1659bd5118d service nova] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1224.655873] env[62692]: DEBUG oslo_vmware.api [None req-00ff7331-4269-47a7-8091-f048ed6ca116 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1142130, 'name': ReconfigVM_Task, 'duration_secs': 0.185494} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.656259] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-00ff7331-4269-47a7-8091-f048ed6ca116 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249091', 'volume_id': '89113b2e-696c-4171-8d9f-479a6c1f41db', 'name': 'volume-89113b2e-696c-4171-8d9f-479a6c1f41db', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '45eccffc-f059-4bef-97da-dacb16f20f88', 'attached_at': '', 'detached_at': '', 'volume_id': '89113b2e-696c-4171-8d9f-479a6c1f41db', 'serial': '89113b2e-696c-4171-8d9f-479a6c1f41db'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1224.694677] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142129, 'name': Rename_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.809108] env[62692]: DEBUG nova.network.neutron [-] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1224.928121] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-014f6f82-cad1-41bc-877f-1ba7fa26049c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.939487] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6e571a2-dce2-4903-aff2-8b8fa4430fb0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.969269] env[62692]: DEBUG nova.compute.manager [req-9fa71750-13e9-4716-ab16-20193e4594f3 req-b3c2f81e-3715-4ecf-b2fd-b1659bd5118d service nova] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Detach interface failed, port_id=21e76034-6c0f-4938-afda-0d005cf45759, reason: Instance 90932f05-fad3-4c6d-87ae-ab059351b0be could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1225.195322] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142129, 'name': Rename_Task, 'duration_secs': 1.172355} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.195648] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1225.195948] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3c3e8049-da66-4d3e-914c-9cbb85b3f7d4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.202666] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1225.202666] env[62692]: value = "task-1142131" [ 1225.202666] env[62692]: _type = "Task" [ 1225.202666] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.207641] env[62692]: DEBUG nova.objects.instance [None req-00ff7331-4269-47a7-8091-f048ed6ca116 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lazy-loading 'flavor' on Instance uuid 45eccffc-f059-4bef-97da-dacb16f20f88 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1225.213989] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142131, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.311883] env[62692]: INFO nova.compute.manager [-] [instance: 90932f05-fad3-4c6d-87ae-ab059351b0be] Took 1.54 seconds to deallocate network for instance. [ 1225.607359] env[62692]: DEBUG nova.objects.instance [None req-0928d16a-be84-4ce5-99c0-d0bb570f0200 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lazy-loading 'flavor' on Instance uuid 9d3c0909-6207-46b7-b83b-de4d7d626fd7 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1225.715319] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142131, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.819481] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1225.819718] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1225.819912] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1225.846268] env[62692]: INFO nova.scheduler.client.report [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Deleted allocations for instance 90932f05-fad3-4c6d-87ae-ab059351b0be [ 1226.113438] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0928d16a-be84-4ce5-99c0-d0bb570f0200 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "refresh_cache-9d3c0909-6207-46b7-b83b-de4d7d626fd7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1226.113619] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0928d16a-be84-4ce5-99c0-d0bb570f0200 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired lock "refresh_cache-9d3c0909-6207-46b7-b83b-de4d7d626fd7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1226.113754] env[62692]: DEBUG nova.network.neutron [None req-0928d16a-be84-4ce5-99c0-d0bb570f0200 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1226.113952] env[62692]: DEBUG nova.objects.instance [None req-0928d16a-be84-4ce5-99c0-d0bb570f0200 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lazy-loading 'info_cache' on Instance uuid 9d3c0909-6207-46b7-b83b-de4d7d626fd7 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1226.213504] env[62692]: DEBUG oslo_vmware.api [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142131, 'name': PowerOnVM_Task, 'duration_secs': 0.558386} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1226.213792] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1226.214045] env[62692]: INFO nova.compute.manager [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Took 10.11 seconds to spawn the instance on the hypervisor. [ 1226.214237] env[62692]: DEBUG nova.compute.manager [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1226.215015] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b34a25c2-0df4-4571-85ec-5ade1197f186 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.217721] env[62692]: DEBUG oslo_concurrency.lockutils [None req-00ff7331-4269-47a7-8091-f048ed6ca116 tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "45eccffc-f059-4bef-97da-dacb16f20f88" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.257s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1226.243442] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1cae7d4c-a99f-4281-8491-4f850b76328a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "45eccffc-f059-4bef-97da-dacb16f20f88" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1226.243442] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1cae7d4c-a99f-4281-8491-4f850b76328a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "45eccffc-f059-4bef-97da-dacb16f20f88" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1226.356438] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5aced8c5-d3f3-4341-82ad-dadb3b5f2fd6 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "90932f05-fad3-4c6d-87ae-ab059351b0be" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.703s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1226.618167] env[62692]: DEBUG nova.objects.base [None req-0928d16a-be84-4ce5-99c0-d0bb570f0200 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Object Instance<9d3c0909-6207-46b7-b83b-de4d7d626fd7> lazy-loaded attributes: flavor,info_cache {{(pid=62692) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1226.736122] env[62692]: INFO nova.compute.manager [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Took 21.80 seconds to build instance. [ 1226.746143] env[62692]: INFO nova.compute.manager [None req-1cae7d4c-a99f-4281-8491-4f850b76328a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Detaching volume 8314cd71-4b96-4fcc-a78e-814d3eea76e0 [ 1226.777978] env[62692]: INFO nova.virt.block_device [None req-1cae7d4c-a99f-4281-8491-4f850b76328a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Attempting to driver detach volume 8314cd71-4b96-4fcc-a78e-814d3eea76e0 from mountpoint /dev/sdc [ 1226.778259] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-1cae7d4c-a99f-4281-8491-4f850b76328a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Volume detach. Driver type: vmdk {{(pid=62692) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1226.778453] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-1cae7d4c-a99f-4281-8491-4f850b76328a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249094', 'volume_id': '8314cd71-4b96-4fcc-a78e-814d3eea76e0', 'name': 'volume-8314cd71-4b96-4fcc-a78e-814d3eea76e0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '45eccffc-f059-4bef-97da-dacb16f20f88', 'attached_at': '', 'detached_at': '', 'volume_id': '8314cd71-4b96-4fcc-a78e-814d3eea76e0', 'serial': '8314cd71-4b96-4fcc-a78e-814d3eea76e0'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1226.779372] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceb16154-913d-4480-802c-d8db84fce1b0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.801435] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a08f9d6-5abf-4f1c-b40e-fcc35f2f7195 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.809061] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49a44e42-0a31-4990-867e-1185ece4774d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.829933] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f8061a9-481d-4f87-89af-7d1ff04b1982 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.845125] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-1cae7d4c-a99f-4281-8491-4f850b76328a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] The volume has not been displaced from its original location: [datastore2] volume-8314cd71-4b96-4fcc-a78e-814d3eea76e0/volume-8314cd71-4b96-4fcc-a78e-814d3eea76e0.vmdk. No consolidation needed. {{(pid=62692) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1226.850376] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-1cae7d4c-a99f-4281-8491-4f850b76328a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Reconfiguring VM instance instance-0000006f to detach disk 2002 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1226.850760] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dcc3ba7b-e984-4b53-b2ce-66f5c547d92b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.868754] env[62692]: DEBUG oslo_vmware.api [None req-1cae7d4c-a99f-4281-8491-4f850b76328a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1226.868754] env[62692]: value = "task-1142133" [ 1226.868754] env[62692]: _type = "Task" [ 1226.868754] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.878836] env[62692]: DEBUG oslo_vmware.api [None req-1cae7d4c-a99f-4281-8491-4f850b76328a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1142133, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.238708] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2880acfc-bcb5-4b38-b4b5-195637ed4c31 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "0806e8c3-8b51-45e8-aaf8-cbd6f03f5489" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.318s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1227.388076] env[62692]: DEBUG oslo_vmware.api [None req-1cae7d4c-a99f-4281-8491-4f850b76328a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1142133, 'name': ReconfigVM_Task, 'duration_secs': 0.421774} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.391673] env[62692]: DEBUG nova.network.neutron [None req-0928d16a-be84-4ce5-99c0-d0bb570f0200 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Updating instance_info_cache with network_info: [{"id": "ea6ff473-6d45-4425-a9e9-535a999fc3a4", "address": "fa:16:3e:10:57:4d", "network": {"id": "f70727d6-a9aa-44b0-a244-2533b3f96571", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-284105837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1037e5d436bd429391159b87a712f6e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea6ff473-6d", "ovs_interfaceid": "ea6ff473-6d45-4425-a9e9-535a999fc3a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1227.392761] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-1cae7d4c-a99f-4281-8491-4f850b76328a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Reconfigured VM instance instance-0000006f to detach disk 2002 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1227.399989] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b81c264-c049-425a-a402-1e4e286e05ad {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.418085] env[62692]: DEBUG oslo_vmware.api [None req-1cae7d4c-a99f-4281-8491-4f850b76328a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1227.418085] env[62692]: value = "task-1142134" [ 1227.418085] env[62692]: _type = "Task" [ 1227.418085] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.426757] env[62692]: DEBUG oslo_vmware.api [None req-1cae7d4c-a99f-4281-8491-4f850b76328a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1142134, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.737312] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "cb18c2e7-3241-4c9a-b1eb-a147b40530a1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1227.737618] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "cb18c2e7-3241-4c9a-b1eb-a147b40530a1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1227.791771] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "29910848-22f2-4c2e-98fe-cd9e32e97865" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1227.792046] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "29910848-22f2-4c2e-98fe-cd9e32e97865" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1227.792270] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "29910848-22f2-4c2e-98fe-cd9e32e97865-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1227.792465] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "29910848-22f2-4c2e-98fe-cd9e32e97865-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1227.792639] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "29910848-22f2-4c2e-98fe-cd9e32e97865-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1227.794906] env[62692]: INFO nova.compute.manager [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Terminating instance [ 1227.796679] env[62692]: DEBUG nova.compute.manager [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1227.796913] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1227.797774] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f25efe4-ffd2-4b5e-ae4e-fc0e3e230566 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.806435] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1227.806669] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-64d0398d-0fd6-4732-a9d6-500a04b2b29c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.813472] env[62692]: DEBUG oslo_vmware.api [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1227.813472] env[62692]: value = "task-1142135" [ 1227.813472] env[62692]: _type = "Task" [ 1227.813472] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.821926] env[62692]: DEBUG oslo_vmware.api [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142135, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.901161] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "0806e8c3-8b51-45e8-aaf8-cbd6f03f5489" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1227.901161] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "0806e8c3-8b51-45e8-aaf8-cbd6f03f5489" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1227.901475] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "0806e8c3-8b51-45e8-aaf8-cbd6f03f5489-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1227.901555] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "0806e8c3-8b51-45e8-aaf8-cbd6f03f5489-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1227.901749] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "0806e8c3-8b51-45e8-aaf8-cbd6f03f5489-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1227.903828] env[62692]: DEBUG oslo_concurrency.lockutils [None req-0928d16a-be84-4ce5-99c0-d0bb570f0200 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Releasing lock "refresh_cache-9d3c0909-6207-46b7-b83b-de4d7d626fd7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1227.905389] env[62692]: INFO nova.compute.manager [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Terminating instance [ 1227.907336] env[62692]: DEBUG nova.compute.manager [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1227.907540] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1227.908517] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4baa17c3-172b-47c1-b1b2-fbc75cf474e8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.917033] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1227.917298] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-29f9bd99-2173-4c08-a025-da2677134e3b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.928042] env[62692]: DEBUG oslo_vmware.api [None req-1cae7d4c-a99f-4281-8491-4f850b76328a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1142134, 'name': ReconfigVM_Task, 'duration_secs': 0.20227} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.929297] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-1cae7d4c-a99f-4281-8491-4f850b76328a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249094', 'volume_id': '8314cd71-4b96-4fcc-a78e-814d3eea76e0', 'name': 'volume-8314cd71-4b96-4fcc-a78e-814d3eea76e0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '45eccffc-f059-4bef-97da-dacb16f20f88', 'attached_at': '', 'detached_at': '', 'volume_id': '8314cd71-4b96-4fcc-a78e-814d3eea76e0', 'serial': '8314cd71-4b96-4fcc-a78e-814d3eea76e0'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1227.931632] env[62692]: DEBUG oslo_vmware.api [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1227.931632] env[62692]: value = "task-1142136" [ 1227.931632] env[62692]: _type = "Task" [ 1227.931632] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.939887] env[62692]: DEBUG oslo_vmware.api [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142136, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.240493] env[62692]: DEBUG nova.compute.manager [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1228.294599] env[62692]: DEBUG oslo_vmware.rw_handles [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5222e1d7-9686-159c-8e2b-9767be4f09dc/disk-0.vmdk. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1228.295757] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e43c93d-fd6f-428d-bfab-781c5bfc6f17 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.302198] env[62692]: DEBUG oslo_vmware.rw_handles [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5222e1d7-9686-159c-8e2b-9767be4f09dc/disk-0.vmdk is in state: ready. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1228.302372] env[62692]: ERROR oslo_vmware.rw_handles [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5222e1d7-9686-159c-8e2b-9767be4f09dc/disk-0.vmdk due to incomplete transfer. [ 1228.302583] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5800d4e8-504d-4336-9181-07b728d016fa {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.309254] env[62692]: DEBUG oslo_vmware.rw_handles [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5222e1d7-9686-159c-8e2b-9767be4f09dc/disk-0.vmdk. {{(pid=62692) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1228.309452] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Uploaded image a9c559b8-09bf-41ff-8b1b-16f309dd363f to the Glance image server {{(pid=62692) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1228.311950] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Destroying the VM {{(pid=62692) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1228.312251] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-996b709c-44d4-4950-b372-caa341f8ece1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.317700] env[62692]: DEBUG oslo_vmware.api [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1228.317700] env[62692]: value = "task-1142137" [ 1228.317700] env[62692]: _type = "Task" [ 1228.317700] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.323592] env[62692]: DEBUG oslo_vmware.api [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142135, 'name': PowerOffVM_Task, 'duration_secs': 0.190715} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.324116] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1228.324297] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1228.324516] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9f26f974-4ce6-4afb-9acf-ba45890dc27a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.328496] env[62692]: DEBUG oslo_vmware.api [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142137, 'name': Destroy_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.390781] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1228.391038] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1228.391237] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Deleting the datastore file [datastore2] 29910848-22f2-4c2e-98fe-cd9e32e97865 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1228.391507] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-50ccd62d-8c0f-4bf6-a19a-bbb7998c8a83 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.397786] env[62692]: DEBUG oslo_vmware.api [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1228.397786] env[62692]: value = "task-1142139" [ 1228.397786] env[62692]: _type = "Task" [ 1228.397786] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.405548] env[62692]: DEBUG oslo_vmware.api [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142139, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.408529] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-0928d16a-be84-4ce5-99c0-d0bb570f0200 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1228.408772] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1e6a6772-579e-4a6f-adf2-29c1bb3757f1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.414651] env[62692]: DEBUG oslo_vmware.api [None req-0928d16a-be84-4ce5-99c0-d0bb570f0200 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1228.414651] env[62692]: value = "task-1142140" [ 1228.414651] env[62692]: _type = "Task" [ 1228.414651] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.422268] env[62692]: DEBUG oslo_vmware.api [None req-0928d16a-be84-4ce5-99c0-d0bb570f0200 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142140, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.440728] env[62692]: DEBUG oslo_vmware.api [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142136, 'name': PowerOffVM_Task, 'duration_secs': 0.199466} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.441090] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1228.441291] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1228.441546] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c3c5d746-913e-4642-b1c9-23cc35c4599d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.476174] env[62692]: DEBUG nova.objects.instance [None req-1cae7d4c-a99f-4281-8491-4f850b76328a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lazy-loading 'flavor' on Instance uuid 45eccffc-f059-4bef-97da-dacb16f20f88 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1228.507166] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1228.507468] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1228.507776] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Deleting the datastore file [datastore2] 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1228.508141] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-020bbbfc-3a0a-40c9-95b1-7db221f16fd5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.514875] env[62692]: DEBUG oslo_vmware.api [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for the task: (returnval){ [ 1228.514875] env[62692]: value = "task-1142142" [ 1228.514875] env[62692]: _type = "Task" [ 1228.514875] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.523650] env[62692]: DEBUG oslo_vmware.api [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142142, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.765593] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1228.765959] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1228.767413] env[62692]: INFO nova.compute.claims [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1228.828447] env[62692]: DEBUG oslo_vmware.api [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142137, 'name': Destroy_Task} progress is 33%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.908542] env[62692]: DEBUG oslo_vmware.api [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142139, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.183512} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.908875] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1228.909148] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1228.909401] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1228.909638] env[62692]: INFO nova.compute.manager [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1228.909941] env[62692]: DEBUG oslo.service.loopingcall [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1228.910227] env[62692]: DEBUG nova.compute.manager [-] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1228.910321] env[62692]: DEBUG nova.network.neutron [-] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1228.927879] env[62692]: DEBUG oslo_vmware.api [None req-0928d16a-be84-4ce5-99c0-d0bb570f0200 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142140, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.025020] env[62692]: DEBUG oslo_vmware.api [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Task: {'id': task-1142142, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.180221} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.025311] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1229.025590] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1229.025674] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1229.025846] env[62692]: INFO nova.compute.manager [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1229.026098] env[62692]: DEBUG oslo.service.loopingcall [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1229.026336] env[62692]: DEBUG nova.compute.manager [-] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1229.026452] env[62692]: DEBUG nova.network.neutron [-] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1229.145214] env[62692]: DEBUG nova.compute.manager [req-be20b427-db35-44c1-a897-a87587d22e05 req-702d3755-ac7e-416b-9eb1-5533ce7bc74a service nova] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Received event network-vif-deleted-e7e22660-7340-4f0b-8d51-ad4b8834c59b {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1229.145453] env[62692]: INFO nova.compute.manager [req-be20b427-db35-44c1-a897-a87587d22e05 req-702d3755-ac7e-416b-9eb1-5533ce7bc74a service nova] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Neutron deleted interface e7e22660-7340-4f0b-8d51-ad4b8834c59b; detaching it from the instance and deleting it from the info cache [ 1229.145649] env[62692]: DEBUG nova.network.neutron [req-be20b427-db35-44c1-a897-a87587d22e05 req-702d3755-ac7e-416b-9eb1-5533ce7bc74a service nova] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1229.332637] env[62692]: DEBUG oslo_vmware.api [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142137, 'name': Destroy_Task, 'duration_secs': 0.559785} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.333053] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Destroyed the VM [ 1229.333581] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Deleting Snapshot of the VM instance {{(pid=62692) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1229.333966] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-fae39f93-5960-494d-921f-f24beeb391e0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.341781] env[62692]: DEBUG oslo_vmware.api [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1229.341781] env[62692]: value = "task-1142143" [ 1229.341781] env[62692]: _type = "Task" [ 1229.341781] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.354215] env[62692]: DEBUG oslo_vmware.api [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142143, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.425014] env[62692]: DEBUG oslo_vmware.api [None req-0928d16a-be84-4ce5-99c0-d0bb570f0200 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142140, 'name': PowerOnVM_Task, 'duration_secs': 0.549565} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.425681] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-0928d16a-be84-4ce5-99c0-d0bb570f0200 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1229.425924] env[62692]: DEBUG nova.compute.manager [None req-0928d16a-be84-4ce5-99c0-d0bb570f0200 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1229.426904] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5e4512a-bfca-4eb6-a393-71fe820461b4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.483751] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1cae7d4c-a99f-4281-8491-4f850b76328a tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "45eccffc-f059-4bef-97da-dacb16f20f88" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.241s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1229.603059] env[62692]: DEBUG nova.compute.manager [req-ae253dd5-06b2-4c72-b9e3-afbab3baea9f req-563b6e55-0846-41df-bf0c-dc8f76a116ff service nova] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Received event network-vif-deleted-7f32f75c-64a7-4073-9dde-0a4510530607 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1229.603278] env[62692]: INFO nova.compute.manager [req-ae253dd5-06b2-4c72-b9e3-afbab3baea9f req-563b6e55-0846-41df-bf0c-dc8f76a116ff service nova] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Neutron deleted interface 7f32f75c-64a7-4073-9dde-0a4510530607; detaching it from the instance and deleting it from the info cache [ 1229.603461] env[62692]: DEBUG nova.network.neutron [req-ae253dd5-06b2-4c72-b9e3-afbab3baea9f req-563b6e55-0846-41df-bf0c-dc8f76a116ff service nova] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1229.613161] env[62692]: DEBUG nova.network.neutron [-] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1229.648250] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dc994481-78b1-47b2-bd1b-ce666e7b0672 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.659791] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03480338-aa4c-4bff-b46c-c431b0f47ac7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.691868] env[62692]: DEBUG nova.compute.manager [req-be20b427-db35-44c1-a897-a87587d22e05 req-702d3755-ac7e-416b-9eb1-5533ce7bc74a service nova] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Detach interface failed, port_id=e7e22660-7340-4f0b-8d51-ad4b8834c59b, reason: Instance 29910848-22f2-4c2e-98fe-cd9e32e97865 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1229.854860] env[62692]: DEBUG oslo_vmware.api [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142143, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.911930] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e3c67f5-7bf3-4805-a1e0-83d13711e5a9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.919908] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9951e83-11ed-418d-b6e5-de3f025c69f1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.954319] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67381e40-ffda-4183-a08c-d83069c0b90a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.963578] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02f2ce37-d570-4513-9b72-fa2bc21dc280 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.977631] env[62692]: DEBUG nova.compute.provider_tree [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1230.079231] env[62692]: DEBUG nova.network.neutron [-] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1230.107444] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-90ade0b2-5435-414b-a4ee-e06dab61ff70 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.118327] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d925026-007a-47d5-939b-f6a0a45a34cf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.130898] env[62692]: INFO nova.compute.manager [-] [instance: 29910848-22f2-4c2e-98fe-cd9e32e97865] Took 1.22 seconds to deallocate network for instance. [ 1230.154836] env[62692]: DEBUG nova.compute.manager [req-ae253dd5-06b2-4c72-b9e3-afbab3baea9f req-563b6e55-0846-41df-bf0c-dc8f76a116ff service nova] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Detach interface failed, port_id=7f32f75c-64a7-4073-9dde-0a4510530607, reason: Instance 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1230.356610] env[62692]: DEBUG oslo_vmware.api [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142143, 'name': RemoveSnapshot_Task, 'duration_secs': 0.772691} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.356610] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Deleted Snapshot of the VM instance {{(pid=62692) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1230.356610] env[62692]: DEBUG nova.compute.manager [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1230.357387] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8be5fb8-4239-48df-8f3b-a48844f252c9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.481350] env[62692]: DEBUG nova.scheduler.client.report [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1230.582330] env[62692]: INFO nova.compute.manager [-] [instance: 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489] Took 1.56 seconds to deallocate network for instance. [ 1230.644243] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1230.707720] env[62692]: DEBUG oslo_concurrency.lockutils [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "45eccffc-f059-4bef-97da-dacb16f20f88" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1230.708011] env[62692]: DEBUG oslo_concurrency.lockutils [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "45eccffc-f059-4bef-97da-dacb16f20f88" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1230.708252] env[62692]: DEBUG oslo_concurrency.lockutils [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "45eccffc-f059-4bef-97da-dacb16f20f88-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1230.708443] env[62692]: DEBUG oslo_concurrency.lockutils [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "45eccffc-f059-4bef-97da-dacb16f20f88-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1230.708624] env[62692]: DEBUG oslo_concurrency.lockutils [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "45eccffc-f059-4bef-97da-dacb16f20f88-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1230.710985] env[62692]: INFO nova.compute.manager [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Terminating instance [ 1230.713185] env[62692]: DEBUG nova.compute.manager [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1230.713185] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1230.713938] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c7bc26e-208c-43e2-b6e5-de1ec56a1f51 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.722182] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1230.722428] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b54de8cd-99b1-403d-bd90-8cf50b1f5d46 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.729123] env[62692]: DEBUG oslo_vmware.api [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1230.729123] env[62692]: value = "task-1142144" [ 1230.729123] env[62692]: _type = "Task" [ 1230.729123] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.737241] env[62692]: DEBUG oslo_vmware.api [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1142144, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.872107] env[62692]: INFO nova.compute.manager [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Shelve offloading [ 1230.874011] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1230.874358] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c026e00c-33b3-49ec-86c8-4e073f5beec4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.882208] env[62692]: DEBUG oslo_vmware.api [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1230.882208] env[62692]: value = "task-1142145" [ 1230.882208] env[62692]: _type = "Task" [ 1230.882208] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.891886] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] VM already powered off {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1230.892129] env[62692]: DEBUG nova.compute.manager [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1230.893147] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3acbea08-22f6-4e6b-9f65-15d000ddbbc8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.899691] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquiring lock "refresh_cache-8abacc6f-054f-4c4f-abb2-83f20748cf1c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1230.899862] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquired lock "refresh_cache-8abacc6f-054f-4c4f-abb2-83f20748cf1c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1230.900059] env[62692]: DEBUG nova.network.neutron [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1230.986901] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.221s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1230.987532] env[62692]: DEBUG nova.compute.manager [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1230.990237] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.346s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1230.990467] env[62692]: DEBUG nova.objects.instance [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lazy-loading 'resources' on Instance uuid 29910848-22f2-4c2e-98fe-cd9e32e97865 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1231.090940] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1231.152859] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94d55f2e-ac40-4a5a-91d0-d21ef929fba0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.160301] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1700989c-2aac-41a1-8452-88a5eb442c15 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Suspending the VM {{(pid=62692) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1231.160517] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-77c35fd9-d4c3-44c5-91af-e1c5e5b2841a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.166412] env[62692]: DEBUG oslo_vmware.api [None req-1700989c-2aac-41a1-8452-88a5eb442c15 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1231.166412] env[62692]: value = "task-1142146" [ 1231.166412] env[62692]: _type = "Task" [ 1231.166412] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.175908] env[62692]: DEBUG oslo_vmware.api [None req-1700989c-2aac-41a1-8452-88a5eb442c15 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142146, 'name': SuspendVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.239809] env[62692]: DEBUG oslo_vmware.api [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1142144, 'name': PowerOffVM_Task, 'duration_secs': 0.2548} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.240090] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1231.240296] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1231.240614] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f035762b-575c-438a-8a51-08e506a86036 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.306142] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1231.306376] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1231.306620] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Deleting the datastore file [datastore1] 45eccffc-f059-4bef-97da-dacb16f20f88 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1231.306922] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3af76d68-8063-448a-bd21-c1fdfbb72ffc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.314459] env[62692]: DEBUG oslo_vmware.api [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for the task: (returnval){ [ 1231.314459] env[62692]: value = "task-1142148" [ 1231.314459] env[62692]: _type = "Task" [ 1231.314459] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.322503] env[62692]: DEBUG oslo_vmware.api [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1142148, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.494112] env[62692]: DEBUG nova.compute.utils [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1231.496162] env[62692]: DEBUG nova.compute.manager [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1231.496466] env[62692]: DEBUG nova.network.neutron [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1231.556652] env[62692]: DEBUG nova.policy [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6102acb4c19642d8a8190d828f5b5dfc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '697b4281872f4547ad3d051e644b95ce', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1231.611304] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e1fe53b-2170-470e-8ade-f177e793c623 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.621477] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f200462-f662-4976-a6fb-6cbcafab829e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.654779] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ed248a1-3470-4635-988f-a4ee707db033 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.662558] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61cf7710-8dc1-47b3-ab37-4eeba87468a2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.678155] env[62692]: DEBUG nova.compute.provider_tree [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1231.685392] env[62692]: DEBUG oslo_vmware.api [None req-1700989c-2aac-41a1-8452-88a5eb442c15 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142146, 'name': SuspendVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.690043] env[62692]: DEBUG nova.network.neutron [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Updating instance_info_cache with network_info: [{"id": "d463b030-eded-4099-ae6f-2aea2d1af078", "address": "fa:16:3e:db:5f:6a", "network": {"id": "189f2464-2a0a-4637-b0f3-2f6ff8fdb324", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1462324860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f94f7500bf524e3890734a58937b80c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd463b030-ed", "ovs_interfaceid": "d463b030-eded-4099-ae6f-2aea2d1af078", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1231.827674] env[62692]: DEBUG oslo_vmware.api [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Task: {'id': task-1142148, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.256251} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.831545] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1231.831759] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1231.831940] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1231.832145] env[62692]: INFO nova.compute.manager [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1231.832402] env[62692]: DEBUG oslo.service.loopingcall [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1231.832982] env[62692]: DEBUG nova.compute.manager [-] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1231.833108] env[62692]: DEBUG nova.network.neutron [-] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1231.892028] env[62692]: DEBUG nova.network.neutron [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Successfully created port: f84cd756-9d8a-4da6-bcd3-c377d3a0a875 {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1232.003597] env[62692]: DEBUG nova.compute.manager [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1232.177937] env[62692]: DEBUG oslo_vmware.api [None req-1700989c-2aac-41a1-8452-88a5eb442c15 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142146, 'name': SuspendVM_Task, 'duration_secs': 0.628637} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.178259] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1700989c-2aac-41a1-8452-88a5eb442c15 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Suspended the VM {{(pid=62692) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1232.178416] env[62692]: DEBUG nova.compute.manager [None req-1700989c-2aac-41a1-8452-88a5eb442c15 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1232.179181] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e62f80c-2e8b-40ed-976b-3081b06debc2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.182208] env[62692]: DEBUG nova.scheduler.client.report [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1232.193197] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Releasing lock "refresh_cache-8abacc6f-054f-4c4f-abb2-83f20748cf1c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1232.353862] env[62692]: DEBUG nova.compute.manager [req-3120898f-abc8-4aa4-9426-a960b33c4787 req-fcaac539-359e-401d-a5ce-402cea5b47af service nova] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Received event network-vif-deleted-01bf6621-08ff-4983-8bcb-1d178bda2c60 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1232.354394] env[62692]: INFO nova.compute.manager [req-3120898f-abc8-4aa4-9426-a960b33c4787 req-fcaac539-359e-401d-a5ce-402cea5b47af service nova] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Neutron deleted interface 01bf6621-08ff-4983-8bcb-1d178bda2c60; detaching it from the instance and deleting it from the info cache [ 1232.354522] env[62692]: DEBUG nova.network.neutron [req-3120898f-abc8-4aa4-9426-a960b33c4787 req-fcaac539-359e-401d-a5ce-402cea5b47af service nova] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1232.544593] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1232.545631] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc2d2462-5712-4d00-b0db-79a6d894cb86 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.553823] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1232.554110] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0b84dc49-4735-45e0-b186-533ad9898e76 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.670666] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1232.670903] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1232.671119] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Deleting the datastore file [datastore1] 8abacc6f-054f-4c4f-abb2-83f20748cf1c {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1232.671392] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3ad35ebb-ca9c-42bb-adb2-bae58a8042ba {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.677658] env[62692]: DEBUG oslo_vmware.api [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1232.677658] env[62692]: value = "task-1142150" [ 1232.677658] env[62692]: _type = "Task" [ 1232.677658] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.685750] env[62692]: DEBUG oslo_vmware.api [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142150, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.687571] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.697s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1232.689792] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.599s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1232.690113] env[62692]: DEBUG nova.objects.instance [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lazy-loading 'resources' on Instance uuid 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1232.715638] env[62692]: INFO nova.scheduler.client.report [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Deleted allocations for instance 29910848-22f2-4c2e-98fe-cd9e32e97865 [ 1232.824729] env[62692]: DEBUG nova.network.neutron [-] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1232.856671] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d2189e06-a87e-4373-ac3c-9870b7a5bb47 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.867255] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f70d204d-904d-4b0d-9334-57ae4ccde598 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.895204] env[62692]: DEBUG nova.compute.manager [req-3120898f-abc8-4aa4-9426-a960b33c4787 req-fcaac539-359e-401d-a5ce-402cea5b47af service nova] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Detach interface failed, port_id=01bf6621-08ff-4983-8bcb-1d178bda2c60, reason: Instance 45eccffc-f059-4bef-97da-dacb16f20f88 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1233.019079] env[62692]: DEBUG nova.compute.manager [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1233.120768] env[62692]: DEBUG nova.virt.hardware [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1233.120768] env[62692]: DEBUG nova.virt.hardware [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1233.120768] env[62692]: DEBUG nova.virt.hardware [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1233.120768] env[62692]: DEBUG nova.virt.hardware [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1233.120768] env[62692]: DEBUG nova.virt.hardware [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1233.120768] env[62692]: DEBUG nova.virt.hardware [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1233.120768] env[62692]: DEBUG nova.virt.hardware [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1233.120768] env[62692]: DEBUG nova.virt.hardware [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1233.120768] env[62692]: DEBUG nova.virt.hardware [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1233.120768] env[62692]: DEBUG nova.virt.hardware [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1233.120768] env[62692]: DEBUG nova.virt.hardware [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1233.120768] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84fccf2b-755d-4919-ab65-fc0179db9978 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.127386] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7981047a-0d31-4f2f-998c-cfd004bdd8c5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.187588] env[62692]: DEBUG oslo_vmware.api [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142150, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147147} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.187859] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1233.188068] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1233.188260] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1233.210896] env[62692]: INFO nova.scheduler.client.report [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Deleted allocations for instance 8abacc6f-054f-4c4f-abb2-83f20748cf1c [ 1233.222866] env[62692]: DEBUG oslo_concurrency.lockutils [None req-5d189cc0-8ccf-4da7-8eba-719496c4dea4 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "29910848-22f2-4c2e-98fe-cd9e32e97865" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.431s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1233.274786] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-953dffc7-ee44-4f97-b521-611916162ae8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.282951] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0939b19c-69b0-4ceb-9c39-1bbbfa9abdd2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.314807] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1eaff82-ff52-4600-8eb6-1547a978ce9a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.321904] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4ebd62b-e596-4a9a-9651-61258df06b88 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.327323] env[62692]: INFO nova.compute.manager [-] [instance: 45eccffc-f059-4bef-97da-dacb16f20f88] Took 1.49 seconds to deallocate network for instance. [ 1233.336465] env[62692]: DEBUG nova.compute.provider_tree [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1233.387826] env[62692]: DEBUG nova.compute.manager [req-8359fe3e-012f-4aa0-9865-dd5fb869ad7a req-85615f1b-38a5-4ba9-811e-a220612965e8 service nova] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Received event network-vif-plugged-f84cd756-9d8a-4da6-bcd3-c377d3a0a875 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1233.388071] env[62692]: DEBUG oslo_concurrency.lockutils [req-8359fe3e-012f-4aa0-9865-dd5fb869ad7a req-85615f1b-38a5-4ba9-811e-a220612965e8 service nova] Acquiring lock "cb18c2e7-3241-4c9a-b1eb-a147b40530a1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1233.388324] env[62692]: DEBUG oslo_concurrency.lockutils [req-8359fe3e-012f-4aa0-9865-dd5fb869ad7a req-85615f1b-38a5-4ba9-811e-a220612965e8 service nova] Lock "cb18c2e7-3241-4c9a-b1eb-a147b40530a1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1233.388452] env[62692]: DEBUG oslo_concurrency.lockutils [req-8359fe3e-012f-4aa0-9865-dd5fb869ad7a req-85615f1b-38a5-4ba9-811e-a220612965e8 service nova] Lock "cb18c2e7-3241-4c9a-b1eb-a147b40530a1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1233.388622] env[62692]: DEBUG nova.compute.manager [req-8359fe3e-012f-4aa0-9865-dd5fb869ad7a req-85615f1b-38a5-4ba9-811e-a220612965e8 service nova] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] No waiting events found dispatching network-vif-plugged-f84cd756-9d8a-4da6-bcd3-c377d3a0a875 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1233.388787] env[62692]: WARNING nova.compute.manager [req-8359fe3e-012f-4aa0-9865-dd5fb869ad7a req-85615f1b-38a5-4ba9-811e-a220612965e8 service nova] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Received unexpected event network-vif-plugged-f84cd756-9d8a-4da6-bcd3-c377d3a0a875 for instance with vm_state building and task_state spawning. [ 1233.478099] env[62692]: DEBUG nova.network.neutron [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Successfully updated port: f84cd756-9d8a-4da6-bcd3-c377d3a0a875 {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1233.652071] env[62692]: INFO nova.compute.manager [None req-ed48f178-0fee-4f81-937a-e27448fbae27 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Resuming [ 1233.653104] env[62692]: DEBUG nova.objects.instance [None req-ed48f178-0fee-4f81-937a-e27448fbae27 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lazy-loading 'flavor' on Instance uuid 9d3c0909-6207-46b7-b83b-de4d7d626fd7 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1233.715227] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1233.841915] env[62692]: DEBUG nova.scheduler.client.report [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1233.845661] env[62692]: DEBUG oslo_concurrency.lockutils [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1233.980097] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "refresh_cache-cb18c2e7-3241-4c9a-b1eb-a147b40530a1" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1233.980423] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquired lock "refresh_cache-cb18c2e7-3241-4c9a-b1eb-a147b40530a1" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1233.980423] env[62692]: DEBUG nova.network.neutron [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1234.347200] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.657s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1234.349616] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.634s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1234.349892] env[62692]: DEBUG nova.objects.instance [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lazy-loading 'resources' on Instance uuid 8abacc6f-054f-4c4f-abb2-83f20748cf1c {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1234.364825] env[62692]: INFO nova.scheduler.client.report [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Deleted allocations for instance 0806e8c3-8b51-45e8-aaf8-cbd6f03f5489 [ 1234.398887] env[62692]: DEBUG nova.compute.manager [req-0082cda9-65bf-47d4-b139-ffa7cce7e326 req-d3798151-6862-433e-ac9a-8f98e266657c service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Received event network-vif-unplugged-d463b030-eded-4099-ae6f-2aea2d1af078 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1234.399184] env[62692]: DEBUG oslo_concurrency.lockutils [req-0082cda9-65bf-47d4-b139-ffa7cce7e326 req-d3798151-6862-433e-ac9a-8f98e266657c service nova] Acquiring lock "8abacc6f-054f-4c4f-abb2-83f20748cf1c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1234.399364] env[62692]: DEBUG oslo_concurrency.lockutils [req-0082cda9-65bf-47d4-b139-ffa7cce7e326 req-d3798151-6862-433e-ac9a-8f98e266657c service nova] Lock "8abacc6f-054f-4c4f-abb2-83f20748cf1c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1234.399544] env[62692]: DEBUG oslo_concurrency.lockutils [req-0082cda9-65bf-47d4-b139-ffa7cce7e326 req-d3798151-6862-433e-ac9a-8f98e266657c service nova] Lock "8abacc6f-054f-4c4f-abb2-83f20748cf1c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1234.399719] env[62692]: DEBUG nova.compute.manager [req-0082cda9-65bf-47d4-b139-ffa7cce7e326 req-d3798151-6862-433e-ac9a-8f98e266657c service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] No waiting events found dispatching network-vif-unplugged-d463b030-eded-4099-ae6f-2aea2d1af078 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1234.399895] env[62692]: WARNING nova.compute.manager [req-0082cda9-65bf-47d4-b139-ffa7cce7e326 req-d3798151-6862-433e-ac9a-8f98e266657c service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Received unexpected event network-vif-unplugged-d463b030-eded-4099-ae6f-2aea2d1af078 for instance with vm_state shelved_offloaded and task_state None. [ 1234.400073] env[62692]: DEBUG nova.compute.manager [req-0082cda9-65bf-47d4-b139-ffa7cce7e326 req-d3798151-6862-433e-ac9a-8f98e266657c service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Received event network-changed-d463b030-eded-4099-ae6f-2aea2d1af078 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1234.400235] env[62692]: DEBUG nova.compute.manager [req-0082cda9-65bf-47d4-b139-ffa7cce7e326 req-d3798151-6862-433e-ac9a-8f98e266657c service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Refreshing instance network info cache due to event network-changed-d463b030-eded-4099-ae6f-2aea2d1af078. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1234.401562] env[62692]: DEBUG oslo_concurrency.lockutils [req-0082cda9-65bf-47d4-b139-ffa7cce7e326 req-d3798151-6862-433e-ac9a-8f98e266657c service nova] Acquiring lock "refresh_cache-8abacc6f-054f-4c4f-abb2-83f20748cf1c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1234.401562] env[62692]: DEBUG oslo_concurrency.lockutils [req-0082cda9-65bf-47d4-b139-ffa7cce7e326 req-d3798151-6862-433e-ac9a-8f98e266657c service nova] Acquired lock "refresh_cache-8abacc6f-054f-4c4f-abb2-83f20748cf1c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1234.401562] env[62692]: DEBUG nova.network.neutron [req-0082cda9-65bf-47d4-b139-ffa7cce7e326 req-d3798151-6862-433e-ac9a-8f98e266657c service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Refreshing network info cache for port d463b030-eded-4099-ae6f-2aea2d1af078 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1234.512917] env[62692]: DEBUG nova.network.neutron [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1234.647520] env[62692]: DEBUG nova.network.neutron [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Updating instance_info_cache with network_info: [{"id": "f84cd756-9d8a-4da6-bcd3-c377d3a0a875", "address": "fa:16:3e:88:fe:19", "network": {"id": "77695f27-3027-4a92-8b9b-0a90cdc53f62", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-337681783-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697b4281872f4547ad3d051e644b95ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf84cd756-9d", "ovs_interfaceid": "f84cd756-9d8a-4da6-bcd3-c377d3a0a875", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1234.660461] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ed48f178-0fee-4f81-937a-e27448fbae27 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "refresh_cache-9d3c0909-6207-46b7-b83b-de4d7d626fd7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1234.660785] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ed48f178-0fee-4f81-937a-e27448fbae27 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquired lock "refresh_cache-9d3c0909-6207-46b7-b83b-de4d7d626fd7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1234.661012] env[62692]: DEBUG nova.network.neutron [None req-ed48f178-0fee-4f81-937a-e27448fbae27 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1234.855492] env[62692]: DEBUG nova.objects.instance [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lazy-loading 'numa_topology' on Instance uuid 8abacc6f-054f-4c4f-abb2-83f20748cf1c {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1234.872061] env[62692]: DEBUG oslo_concurrency.lockutils [None req-2907062c-c216-4dca-977f-f3f2f565a3a9 tempest-MultipleCreateTestJSON-1500388774 tempest-MultipleCreateTestJSON-1500388774-project-member] Lock "0806e8c3-8b51-45e8-aaf8-cbd6f03f5489" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.971s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1235.090327] env[62692]: DEBUG nova.network.neutron [req-0082cda9-65bf-47d4-b139-ffa7cce7e326 req-d3798151-6862-433e-ac9a-8f98e266657c service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Updated VIF entry in instance network info cache for port d463b030-eded-4099-ae6f-2aea2d1af078. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1235.090695] env[62692]: DEBUG nova.network.neutron [req-0082cda9-65bf-47d4-b139-ffa7cce7e326 req-d3798151-6862-433e-ac9a-8f98e266657c service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Updating instance_info_cache with network_info: [{"id": "d463b030-eded-4099-ae6f-2aea2d1af078", "address": "fa:16:3e:db:5f:6a", "network": {"id": "189f2464-2a0a-4637-b0f3-2f6ff8fdb324", "bridge": null, "label": "tempest-ServersNegativeTestJSON-1462324860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f94f7500bf524e3890734a58937b80c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapd463b030-ed", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1235.150040] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Releasing lock "refresh_cache-cb18c2e7-3241-4c9a-b1eb-a147b40530a1" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1235.150333] env[62692]: DEBUG nova.compute.manager [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Instance network_info: |[{"id": "f84cd756-9d8a-4da6-bcd3-c377d3a0a875", "address": "fa:16:3e:88:fe:19", "network": {"id": "77695f27-3027-4a92-8b9b-0a90cdc53f62", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-337681783-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697b4281872f4547ad3d051e644b95ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf84cd756-9d", "ovs_interfaceid": "f84cd756-9d8a-4da6-bcd3-c377d3a0a875", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1235.150749] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:88:fe:19', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8e7f6f41-f4eb-4832-a390-730fca1cf717', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f84cd756-9d8a-4da6-bcd3-c377d3a0a875', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1235.158152] env[62692]: DEBUG oslo.service.loopingcall [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1235.158366] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1235.158582] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d6a308ba-0a31-478f-bd4d-543444f80ecf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.182096] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1235.182096] env[62692]: value = "task-1142151" [ 1235.182096] env[62692]: _type = "Task" [ 1235.182096] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.189746] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1142151, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.357980] env[62692]: DEBUG nova.objects.base [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Object Instance<8abacc6f-054f-4c4f-abb2-83f20748cf1c> lazy-loaded attributes: resources,numa_topology {{(pid=62692) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1235.389640] env[62692]: DEBUG nova.network.neutron [None req-ed48f178-0fee-4f81-937a-e27448fbae27 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Updating instance_info_cache with network_info: [{"id": "ea6ff473-6d45-4425-a9e9-535a999fc3a4", "address": "fa:16:3e:10:57:4d", "network": {"id": "f70727d6-a9aa-44b0-a244-2533b3f96571", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-284105837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1037e5d436bd429391159b87a712f6e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f44b2fa3-6730-4b87-8839-947eff21213f", "external-id": "nsx-vlan-transportzone-984", "segmentation_id": 984, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea6ff473-6d", "ovs_interfaceid": "ea6ff473-6d45-4425-a9e9-535a999fc3a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1235.417718] env[62692]: DEBUG nova.compute.manager [req-4a30801c-4a1e-4fc6-b752-de9e4f419090 req-a59aa1ae-cfe5-4e8c-9eba-c9bd5ca93dd5 service nova] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Received event network-changed-f84cd756-9d8a-4da6-bcd3-c377d3a0a875 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1235.417919] env[62692]: DEBUG nova.compute.manager [req-4a30801c-4a1e-4fc6-b752-de9e4f419090 req-a59aa1ae-cfe5-4e8c-9eba-c9bd5ca93dd5 service nova] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Refreshing instance network info cache due to event network-changed-f84cd756-9d8a-4da6-bcd3-c377d3a0a875. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1235.418179] env[62692]: DEBUG oslo_concurrency.lockutils [req-4a30801c-4a1e-4fc6-b752-de9e4f419090 req-a59aa1ae-cfe5-4e8c-9eba-c9bd5ca93dd5 service nova] Acquiring lock "refresh_cache-cb18c2e7-3241-4c9a-b1eb-a147b40530a1" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1235.418254] env[62692]: DEBUG oslo_concurrency.lockutils [req-4a30801c-4a1e-4fc6-b752-de9e4f419090 req-a59aa1ae-cfe5-4e8c-9eba-c9bd5ca93dd5 service nova] Acquired lock "refresh_cache-cb18c2e7-3241-4c9a-b1eb-a147b40530a1" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1235.418422] env[62692]: DEBUG nova.network.neutron [req-4a30801c-4a1e-4fc6-b752-de9e4f419090 req-a59aa1ae-cfe5-4e8c-9eba-c9bd5ca93dd5 service nova] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Refreshing network info cache for port f84cd756-9d8a-4da6-bcd3-c377d3a0a875 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1235.449420] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af978b44-cad9-47dd-a941-6331dfdcb844 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.457840] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd12439c-4b5b-402c-a3a0-f818abed342a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.491463] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37bec6c8-d35d-4802-bec5-09bb020e403f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.499313] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63b1af7e-1da5-4055-9678-18d03e0dc5ef {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.512984] env[62692]: DEBUG nova.compute.provider_tree [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1235.593935] env[62692]: DEBUG oslo_concurrency.lockutils [req-0082cda9-65bf-47d4-b139-ffa7cce7e326 req-d3798151-6862-433e-ac9a-8f98e266657c service nova] Releasing lock "refresh_cache-8abacc6f-054f-4c4f-abb2-83f20748cf1c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1235.695022] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1142151, 'name': CreateVM_Task, 'duration_secs': 0.29986} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.695022] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1235.695022] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1235.695022] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1235.695022] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1235.695022] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0a1b87c-b85a-46a5-89a7-b158596d1f93 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.698879] env[62692]: DEBUG oslo_vmware.api [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1235.698879] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52e3752f-1006-5f7b-d490-6665b17db17f" [ 1235.698879] env[62692]: _type = "Task" [ 1235.698879] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.706358] env[62692]: DEBUG oslo_vmware.api [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52e3752f-1006-5f7b-d490-6665b17db17f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.749154] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquiring lock "8abacc6f-054f-4c4f-abb2-83f20748cf1c" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1235.892969] env[62692]: DEBUG oslo_concurrency.lockutils [None req-ed48f178-0fee-4f81-937a-e27448fbae27 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Releasing lock "refresh_cache-9d3c0909-6207-46b7-b83b-de4d7d626fd7" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1235.894030] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0296e2ab-b8cf-4670-8658-90453144dd70 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.901198] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ed48f178-0fee-4f81-937a-e27448fbae27 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Resuming the VM {{(pid=62692) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 1235.901198] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a87ca6d3-f776-44ca-bb76-b9c07ed008d6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.908015] env[62692]: DEBUG oslo_vmware.api [None req-ed48f178-0fee-4f81-937a-e27448fbae27 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1235.908015] env[62692]: value = "task-1142152" [ 1235.908015] env[62692]: _type = "Task" [ 1235.908015] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.915372] env[62692]: DEBUG oslo_vmware.api [None req-ed48f178-0fee-4f81-937a-e27448fbae27 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142152, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.016374] env[62692]: DEBUG nova.scheduler.client.report [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1236.209196] env[62692]: DEBUG oslo_vmware.api [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52e3752f-1006-5f7b-d490-6665b17db17f, 'name': SearchDatastore_Task, 'duration_secs': 0.010514} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.209540] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1236.209791] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1236.210041] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1236.210199] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquired lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1236.210398] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1236.210716] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d89b87fd-1381-4514-9571-fccf24f1a281 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.219457] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1236.219653] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1236.220476] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ebf4126-8a44-4f3b-bdc3-370c76b87b44 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.225638] env[62692]: DEBUG oslo_vmware.api [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1236.225638] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52e6a929-9b1f-8d73-e2ed-483acb2844a0" [ 1236.225638] env[62692]: _type = "Task" [ 1236.225638] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.235603] env[62692]: DEBUG oslo_vmware.api [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52e6a929-9b1f-8d73-e2ed-483acb2844a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.328272] env[62692]: DEBUG nova.network.neutron [req-4a30801c-4a1e-4fc6-b752-de9e4f419090 req-a59aa1ae-cfe5-4e8c-9eba-c9bd5ca93dd5 service nova] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Updated VIF entry in instance network info cache for port f84cd756-9d8a-4da6-bcd3-c377d3a0a875. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1236.328687] env[62692]: DEBUG nova.network.neutron [req-4a30801c-4a1e-4fc6-b752-de9e4f419090 req-a59aa1ae-cfe5-4e8c-9eba-c9bd5ca93dd5 service nova] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Updating instance_info_cache with network_info: [{"id": "f84cd756-9d8a-4da6-bcd3-c377d3a0a875", "address": "fa:16:3e:88:fe:19", "network": {"id": "77695f27-3027-4a92-8b9b-0a90cdc53f62", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-337681783-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697b4281872f4547ad3d051e644b95ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf84cd756-9d", "ovs_interfaceid": "f84cd756-9d8a-4da6-bcd3-c377d3a0a875", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1236.418822] env[62692]: DEBUG oslo_vmware.api [None req-ed48f178-0fee-4f81-937a-e27448fbae27 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142152, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.521654] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.172s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1236.524671] env[62692]: DEBUG oslo_concurrency.lockutils [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.679s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1236.524900] env[62692]: DEBUG nova.objects.instance [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lazy-loading 'resources' on Instance uuid 45eccffc-f059-4bef-97da-dacb16f20f88 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1236.736376] env[62692]: DEBUG oslo_vmware.api [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52e6a929-9b1f-8d73-e2ed-483acb2844a0, 'name': SearchDatastore_Task, 'duration_secs': 0.009976} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.737214] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-829331b2-b202-4b63-bf93-0b1e551104fe {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.742582] env[62692]: DEBUG oslo_vmware.api [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1236.742582] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]525343e7-e6ea-19ed-ad18-a104d54ccbd3" [ 1236.742582] env[62692]: _type = "Task" [ 1236.742582] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.751414] env[62692]: DEBUG oslo_vmware.api [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]525343e7-e6ea-19ed-ad18-a104d54ccbd3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.831818] env[62692]: DEBUG oslo_concurrency.lockutils [req-4a30801c-4a1e-4fc6-b752-de9e4f419090 req-a59aa1ae-cfe5-4e8c-9eba-c9bd5ca93dd5 service nova] Releasing lock "refresh_cache-cb18c2e7-3241-4c9a-b1eb-a147b40530a1" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1236.919195] env[62692]: DEBUG oslo_vmware.api [None req-ed48f178-0fee-4f81-937a-e27448fbae27 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142152, 'name': PowerOnVM_Task, 'duration_secs': 0.533846} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.919570] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-ed48f178-0fee-4f81-937a-e27448fbae27 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Resumed the VM {{(pid=62692) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 1236.919784] env[62692]: DEBUG nova.compute.manager [None req-ed48f178-0fee-4f81-937a-e27448fbae27 tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1236.920604] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a9a72b2-09ec-4bd3-a836-7724292e0004 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.033297] env[62692]: DEBUG oslo_concurrency.lockutils [None req-3527d9f7-92d5-4540-a818-f5d4235103a8 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "8abacc6f-054f-4c4f-abb2-83f20748cf1c" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 24.594s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1237.034595] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "8abacc6f-054f-4c4f-abb2-83f20748cf1c" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.286s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1237.034784] env[62692]: INFO nova.compute.manager [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Unshelving [ 1237.109593] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d24e882-6b7e-463f-aae5-c5c52cb82683 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.117544] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e651dbf-b317-44b3-8c4b-02e90c93b4e1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.148791] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3af9b610-5391-4b29-bc24-f7e7ae925fde {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.156629] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6cd6475-0ec5-4286-8c52-5cdade36a2f1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.173038] env[62692]: DEBUG nova.compute.provider_tree [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1237.254157] env[62692]: DEBUG oslo_vmware.api [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]525343e7-e6ea-19ed-ad18-a104d54ccbd3, 'name': SearchDatastore_Task, 'duration_secs': 0.009824} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1237.254460] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Releasing lock "[datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1237.254713] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] cb18c2e7-3241-4c9a-b1eb-a147b40530a1/cb18c2e7-3241-4c9a-b1eb-a147b40530a1.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1237.254999] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6cdf1184-3538-4a7c-924e-5ee66f6cdc49 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.261430] env[62692]: DEBUG oslo_vmware.api [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1237.261430] env[62692]: value = "task-1142153" [ 1237.261430] env[62692]: _type = "Task" [ 1237.261430] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1237.269730] env[62692]: DEBUG oslo_vmware.api [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142153, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.676353] env[62692]: DEBUG nova.scheduler.client.report [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1237.770833] env[62692]: DEBUG oslo_vmware.api [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142153, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478362} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1237.771154] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore1] cb18c2e7-3241-4c9a-b1eb-a147b40530a1/cb18c2e7-3241-4c9a-b1eb-a147b40530a1.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1237.771382] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1237.771631] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a25c0c8c-991d-47b0-a9c2-4cd9ec1ae9c2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.777590] env[62692]: DEBUG oslo_vmware.api [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1237.777590] env[62692]: value = "task-1142154" [ 1237.777590] env[62692]: _type = "Task" [ 1237.777590] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1237.784588] env[62692]: DEBUG oslo_vmware.api [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142154, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.055450] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1238.182041] env[62692]: DEBUG oslo_concurrency.lockutils [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.657s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1238.184339] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.129s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1238.184577] env[62692]: DEBUG nova.objects.instance [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lazy-loading 'pci_requests' on Instance uuid 8abacc6f-054f-4c4f-abb2-83f20748cf1c {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1238.200604] env[62692]: INFO nova.scheduler.client.report [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Deleted allocations for instance 45eccffc-f059-4bef-97da-dacb16f20f88 [ 1238.287690] env[62692]: DEBUG oslo_vmware.api [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142154, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076358} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.287993] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1238.288809] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2c85ed6-e517-43da-8157-84811edd68e1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.311168] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Reconfiguring VM instance instance-00000078 to attach disk [datastore1] cb18c2e7-3241-4c9a-b1eb-a147b40530a1/cb18c2e7-3241-4c9a-b1eb-a147b40530a1.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1238.311855] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d99d2317-c860-41b8-898f-586da7779a97 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.331802] env[62692]: DEBUG oslo_vmware.api [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1238.331802] env[62692]: value = "task-1142155" [ 1238.331802] env[62692]: _type = "Task" [ 1238.331802] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.339339] env[62692]: DEBUG oslo_vmware.api [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142155, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.501066] env[62692]: DEBUG oslo_concurrency.lockutils [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "9d3c0909-6207-46b7-b83b-de4d7d626fd7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1238.501066] env[62692]: DEBUG oslo_concurrency.lockutils [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "9d3c0909-6207-46b7-b83b-de4d7d626fd7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1238.501288] env[62692]: DEBUG oslo_concurrency.lockutils [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "9d3c0909-6207-46b7-b83b-de4d7d626fd7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1238.501360] env[62692]: DEBUG oslo_concurrency.lockutils [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "9d3c0909-6207-46b7-b83b-de4d7d626fd7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1238.501538] env[62692]: DEBUG oslo_concurrency.lockutils [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "9d3c0909-6207-46b7-b83b-de4d7d626fd7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1238.503776] env[62692]: INFO nova.compute.manager [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Terminating instance [ 1238.505742] env[62692]: DEBUG nova.compute.manager [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1238.506046] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1238.506879] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55dc6319-eafc-445a-aef2-9dd7746c7c72 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.514527] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1238.514750] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-43edf11c-d117-4e6a-a376-e07086ffc035 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.520708] env[62692]: DEBUG oslo_vmware.api [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1238.520708] env[62692]: value = "task-1142156" [ 1238.520708] env[62692]: _type = "Task" [ 1238.520708] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.529653] env[62692]: DEBUG oslo_vmware.api [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142156, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.689305] env[62692]: DEBUG nova.objects.instance [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lazy-loading 'numa_topology' on Instance uuid 8abacc6f-054f-4c4f-abb2-83f20748cf1c {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1238.707444] env[62692]: DEBUG oslo_concurrency.lockutils [None req-88b82265-d900-411c-ab09-c56b1bdd4c5c tempest-AttachVolumeTestJSON-1192065763 tempest-AttachVolumeTestJSON-1192065763-project-member] Lock "45eccffc-f059-4bef-97da-dacb16f20f88" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.999s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1238.841933] env[62692]: DEBUG oslo_vmware.api [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142155, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.030741] env[62692]: DEBUG oslo_vmware.api [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142156, 'name': PowerOffVM_Task, 'duration_secs': 0.351401} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.031070] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1239.031253] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1239.031507] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fd0b6a3d-557e-4bd7-b351-dec538c92ccb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.098977] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1239.099180] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1239.099377] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Deleting the datastore file [datastore2] 9d3c0909-6207-46b7-b83b-de4d7d626fd7 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1239.099655] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e364015d-7ff8-4308-8f6a-e25d76b288c6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.106379] env[62692]: DEBUG oslo_vmware.api [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for the task: (returnval){ [ 1239.106379] env[62692]: value = "task-1142158" [ 1239.106379] env[62692]: _type = "Task" [ 1239.106379] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.114930] env[62692]: DEBUG oslo_vmware.api [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142158, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.192431] env[62692]: INFO nova.compute.claims [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1239.341809] env[62692]: DEBUG oslo_vmware.api [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142155, 'name': ReconfigVM_Task, 'duration_secs': 0.777231} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.342230] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Reconfigured VM instance instance-00000078 to attach disk [datastore1] cb18c2e7-3241-4c9a-b1eb-a147b40530a1/cb18c2e7-3241-4c9a-b1eb-a147b40530a1.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1239.342859] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f0f61524-4084-41c1-b99b-c695ff366a87 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.349277] env[62692]: DEBUG oslo_vmware.api [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1239.349277] env[62692]: value = "task-1142159" [ 1239.349277] env[62692]: _type = "Task" [ 1239.349277] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.356866] env[62692]: DEBUG oslo_vmware.api [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142159, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.616787] env[62692]: DEBUG oslo_vmware.api [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Task: {'id': task-1142158, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.17713} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.618054] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1239.618054] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1239.618054] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1239.618054] env[62692]: INFO nova.compute.manager [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1239.618329] env[62692]: DEBUG oslo.service.loopingcall [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1239.618329] env[62692]: DEBUG nova.compute.manager [-] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1239.618391] env[62692]: DEBUG nova.network.neutron [-] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1239.861190] env[62692]: DEBUG oslo_vmware.api [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142159, 'name': Rename_Task, 'duration_secs': 0.146359} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.861482] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1239.861732] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1fd4af55-a923-4838-b111-e0cb8a1ae9a0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.868256] env[62692]: DEBUG oslo_vmware.api [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1239.868256] env[62692]: value = "task-1142161" [ 1239.868256] env[62692]: _type = "Task" [ 1239.868256] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.875624] env[62692]: DEBUG oslo_vmware.api [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142161, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.120859] env[62692]: DEBUG nova.compute.manager [req-049547a4-b4e5-4626-b8c7-e17e883e7947 req-62c90d80-11ca-4100-b627-f69db46b2447 service nova] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Received event network-vif-deleted-ea6ff473-6d45-4425-a9e9-535a999fc3a4 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1240.121071] env[62692]: INFO nova.compute.manager [req-049547a4-b4e5-4626-b8c7-e17e883e7947 req-62c90d80-11ca-4100-b627-f69db46b2447 service nova] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Neutron deleted interface ea6ff473-6d45-4425-a9e9-535a999fc3a4; detaching it from the instance and deleting it from the info cache [ 1240.121224] env[62692]: DEBUG nova.network.neutron [req-049547a4-b4e5-4626-b8c7-e17e883e7947 req-62c90d80-11ca-4100-b627-f69db46b2447 service nova] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1240.292165] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4bd0607-2c97-4787-afaa-d55abfa8d4a2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.300228] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cb991ef-e23a-4736-b332-e91f2f25bde9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.332863] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-533eeddc-5ba0-473c-829a-72104d3f6657 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.344021] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44cba4be-d51a-4627-bb36-465f8b3dd47a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.354433] env[62692]: DEBUG nova.compute.provider_tree [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1240.377514] env[62692]: DEBUG oslo_vmware.api [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142161, 'name': PowerOnVM_Task, 'duration_secs': 0.474127} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.377782] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1240.377983] env[62692]: INFO nova.compute.manager [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Took 7.36 seconds to spawn the instance on the hypervisor. [ 1240.378215] env[62692]: DEBUG nova.compute.manager [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1240.378988] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99c43c97-a3c0-41bc-a2a3-22a14b7172a8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.572847] env[62692]: DEBUG nova.network.neutron [-] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1240.624078] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f7f072de-f1e4-4be6-bf25-03809c7280e1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.635537] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ee6bd00-8250-4270-8f89-27420e4b3d32 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.669241] env[62692]: DEBUG nova.compute.manager [req-049547a4-b4e5-4626-b8c7-e17e883e7947 req-62c90d80-11ca-4100-b627-f69db46b2447 service nova] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Detach interface failed, port_id=ea6ff473-6d45-4425-a9e9-535a999fc3a4, reason: Instance 9d3c0909-6207-46b7-b83b-de4d7d626fd7 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1240.857390] env[62692]: DEBUG nova.scheduler.client.report [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1240.895911] env[62692]: INFO nova.compute.manager [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Took 12.15 seconds to build instance. [ 1241.075836] env[62692]: INFO nova.compute.manager [-] [instance: 9d3c0909-6207-46b7-b83b-de4d7d626fd7] Took 1.46 seconds to deallocate network for instance. [ 1241.363168] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.179s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1241.393065] env[62692]: INFO nova.network.neutron [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Updating port d463b030-eded-4099-ae6f-2aea2d1af078 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1241.397611] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e508f342-bfce-4f05-ab92-4fc52d6e70cf tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "cb18c2e7-3241-4c9a-b1eb-a147b40530a1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.660s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1241.582791] env[62692]: DEBUG oslo_concurrency.lockutils [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1241.583299] env[62692]: DEBUG oslo_concurrency.lockutils [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1241.583581] env[62692]: DEBUG nova.objects.instance [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lazy-loading 'resources' on Instance uuid 9d3c0909-6207-46b7-b83b-de4d7d626fd7 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1242.146868] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d17166e5-15ef-4ac0-9bfa-96b33cfaa1e2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.151974] env[62692]: DEBUG nova.compute.manager [req-ea3fadcd-0295-410f-9b08-3e001f3dff7e req-dd1fce1d-2cb6-4940-9055-d9cacd4c97b7 service nova] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Received event network-changed-f84cd756-9d8a-4da6-bcd3-c377d3a0a875 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1242.152134] env[62692]: DEBUG nova.compute.manager [req-ea3fadcd-0295-410f-9b08-3e001f3dff7e req-dd1fce1d-2cb6-4940-9055-d9cacd4c97b7 service nova] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Refreshing instance network info cache due to event network-changed-f84cd756-9d8a-4da6-bcd3-c377d3a0a875. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1242.152357] env[62692]: DEBUG oslo_concurrency.lockutils [req-ea3fadcd-0295-410f-9b08-3e001f3dff7e req-dd1fce1d-2cb6-4940-9055-d9cacd4c97b7 service nova] Acquiring lock "refresh_cache-cb18c2e7-3241-4c9a-b1eb-a147b40530a1" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1242.152505] env[62692]: DEBUG oslo_concurrency.lockutils [req-ea3fadcd-0295-410f-9b08-3e001f3dff7e req-dd1fce1d-2cb6-4940-9055-d9cacd4c97b7 service nova] Acquired lock "refresh_cache-cb18c2e7-3241-4c9a-b1eb-a147b40530a1" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1242.152701] env[62692]: DEBUG nova.network.neutron [req-ea3fadcd-0295-410f-9b08-3e001f3dff7e req-dd1fce1d-2cb6-4940-9055-d9cacd4c97b7 service nova] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Refreshing network info cache for port f84cd756-9d8a-4da6-bcd3-c377d3a0a875 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1242.156688] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-422a4a84-9212-444c-a694-450dffb3a79b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.186554] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25432780-0712-4b67-9b8c-e72cb1c5465b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.194064] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca266dde-ba29-4e2c-b992-596954b9bc60 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.207105] env[62692]: DEBUG nova.compute.provider_tree [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1242.710595] env[62692]: DEBUG nova.scheduler.client.report [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1242.893794] env[62692]: DEBUG nova.network.neutron [req-ea3fadcd-0295-410f-9b08-3e001f3dff7e req-dd1fce1d-2cb6-4940-9055-d9cacd4c97b7 service nova] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Updated VIF entry in instance network info cache for port f84cd756-9d8a-4da6-bcd3-c377d3a0a875. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1242.894260] env[62692]: DEBUG nova.network.neutron [req-ea3fadcd-0295-410f-9b08-3e001f3dff7e req-dd1fce1d-2cb6-4940-9055-d9cacd4c97b7 service nova] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Updating instance_info_cache with network_info: [{"id": "f84cd756-9d8a-4da6-bcd3-c377d3a0a875", "address": "fa:16:3e:88:fe:19", "network": {"id": "77695f27-3027-4a92-8b9b-0a90cdc53f62", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-337681783-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697b4281872f4547ad3d051e644b95ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf84cd756-9d", "ovs_interfaceid": "f84cd756-9d8a-4da6-bcd3-c377d3a0a875", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1242.920377] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquiring lock "refresh_cache-8abacc6f-054f-4c4f-abb2-83f20748cf1c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1242.920567] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquired lock "refresh_cache-8abacc6f-054f-4c4f-abb2-83f20748cf1c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1242.920756] env[62692]: DEBUG nova.network.neutron [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1243.215147] env[62692]: DEBUG oslo_concurrency.lockutils [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.632s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1243.238421] env[62692]: INFO nova.scheduler.client.report [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Deleted allocations for instance 9d3c0909-6207-46b7-b83b-de4d7d626fd7 [ 1243.397729] env[62692]: DEBUG oslo_concurrency.lockutils [req-ea3fadcd-0295-410f-9b08-3e001f3dff7e req-dd1fce1d-2cb6-4940-9055-d9cacd4c97b7 service nova] Releasing lock "refresh_cache-cb18c2e7-3241-4c9a-b1eb-a147b40530a1" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1243.618498] env[62692]: DEBUG nova.network.neutron [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Updating instance_info_cache with network_info: [{"id": "d463b030-eded-4099-ae6f-2aea2d1af078", "address": "fa:16:3e:db:5f:6a", "network": {"id": "189f2464-2a0a-4637-b0f3-2f6ff8fdb324", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1462324860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f94f7500bf524e3890734a58937b80c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd463b030-ed", "ovs_interfaceid": "d463b030-eded-4099-ae6f-2aea2d1af078", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1243.746273] env[62692]: DEBUG oslo_concurrency.lockutils [None req-961fb9d8-512c-442b-9e52-a5d441f1b18d tempest-ServerActionsTestJSON-366376345 tempest-ServerActionsTestJSON-366376345-project-member] Lock "9d3c0909-6207-46b7-b83b-de4d7d626fd7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.245s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1244.121076] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Releasing lock "refresh_cache-8abacc6f-054f-4c4f-abb2-83f20748cf1c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1244.146383] env[62692]: DEBUG nova.virt.hardware [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='a3d0c183883e52f05ab34d905855990c',container_format='bare',created_at=2024-10-05T20:39:57Z,direct_url=,disk_format='vmdk',id=a9c559b8-09bf-41ff-8b1b-16f309dd363f,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-1726384947-shelved',owner='f94f7500bf524e3890734a58937b80c4',properties=ImageMetaProps,protected=,size=31666688,status='active',tags=,updated_at=2024-10-05T20:40:13Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1244.146664] env[62692]: DEBUG nova.virt.hardware [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1244.146830] env[62692]: DEBUG nova.virt.hardware [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1244.147036] env[62692]: DEBUG nova.virt.hardware [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1244.147202] env[62692]: DEBUG nova.virt.hardware [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1244.147357] env[62692]: DEBUG nova.virt.hardware [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1244.147571] env[62692]: DEBUG nova.virt.hardware [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1244.147737] env[62692]: DEBUG nova.virt.hardware [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1244.147911] env[62692]: DEBUG nova.virt.hardware [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1244.148207] env[62692]: DEBUG nova.virt.hardware [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1244.148914] env[62692]: DEBUG nova.virt.hardware [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1244.149313] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a19d2e23-efa5-4d37-a441-befa343b8e18 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.157303] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc7c198a-0676-41b5-a441-d06af6ce6b10 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.170784] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:db:5f:6a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '669e4919-e0ad-4e23-9f23-4c5f2be0d858', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd463b030-eded-4099-ae6f-2aea2d1af078', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1244.178138] env[62692]: DEBUG oslo.service.loopingcall [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1244.178417] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1244.178638] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6408fac4-d9bf-43de-abce-e039a6c7318f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.196055] env[62692]: DEBUG nova.compute.manager [req-b03a9d8b-5ff6-4cb7-b7ed-3decb04b560f req-261ea049-861f-499d-b041-5e98db28ba49 service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Received event network-vif-plugged-d463b030-eded-4099-ae6f-2aea2d1af078 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1244.196297] env[62692]: DEBUG oslo_concurrency.lockutils [req-b03a9d8b-5ff6-4cb7-b7ed-3decb04b560f req-261ea049-861f-499d-b041-5e98db28ba49 service nova] Acquiring lock "8abacc6f-054f-4c4f-abb2-83f20748cf1c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1244.196515] env[62692]: DEBUG oslo_concurrency.lockutils [req-b03a9d8b-5ff6-4cb7-b7ed-3decb04b560f req-261ea049-861f-499d-b041-5e98db28ba49 service nova] Lock "8abacc6f-054f-4c4f-abb2-83f20748cf1c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1244.196687] env[62692]: DEBUG oslo_concurrency.lockutils [req-b03a9d8b-5ff6-4cb7-b7ed-3decb04b560f req-261ea049-861f-499d-b041-5e98db28ba49 service nova] Lock "8abacc6f-054f-4c4f-abb2-83f20748cf1c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1244.196861] env[62692]: DEBUG nova.compute.manager [req-b03a9d8b-5ff6-4cb7-b7ed-3decb04b560f req-261ea049-861f-499d-b041-5e98db28ba49 service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] No waiting events found dispatching network-vif-plugged-d463b030-eded-4099-ae6f-2aea2d1af078 {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1244.197040] env[62692]: WARNING nova.compute.manager [req-b03a9d8b-5ff6-4cb7-b7ed-3decb04b560f req-261ea049-861f-499d-b041-5e98db28ba49 service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Received unexpected event network-vif-plugged-d463b030-eded-4099-ae6f-2aea2d1af078 for instance with vm_state shelved_offloaded and task_state spawning. [ 1244.197215] env[62692]: DEBUG nova.compute.manager [req-b03a9d8b-5ff6-4cb7-b7ed-3decb04b560f req-261ea049-861f-499d-b041-5e98db28ba49 service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Received event network-changed-d463b030-eded-4099-ae6f-2aea2d1af078 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1244.197376] env[62692]: DEBUG nova.compute.manager [req-b03a9d8b-5ff6-4cb7-b7ed-3decb04b560f req-261ea049-861f-499d-b041-5e98db28ba49 service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Refreshing instance network info cache due to event network-changed-d463b030-eded-4099-ae6f-2aea2d1af078. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1244.197842] env[62692]: DEBUG oslo_concurrency.lockutils [req-b03a9d8b-5ff6-4cb7-b7ed-3decb04b560f req-261ea049-861f-499d-b041-5e98db28ba49 service nova] Acquiring lock "refresh_cache-8abacc6f-054f-4c4f-abb2-83f20748cf1c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1244.197842] env[62692]: DEBUG oslo_concurrency.lockutils [req-b03a9d8b-5ff6-4cb7-b7ed-3decb04b560f req-261ea049-861f-499d-b041-5e98db28ba49 service nova] Acquired lock "refresh_cache-8abacc6f-054f-4c4f-abb2-83f20748cf1c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1244.197842] env[62692]: DEBUG nova.network.neutron [req-b03a9d8b-5ff6-4cb7-b7ed-3decb04b560f req-261ea049-861f-499d-b041-5e98db28ba49 service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Refreshing network info cache for port d463b030-eded-4099-ae6f-2aea2d1af078 {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1244.204167] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1244.204167] env[62692]: value = "task-1142163" [ 1244.204167] env[62692]: _type = "Task" [ 1244.204167] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.212975] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1142163, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.714814] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1142163, 'name': CreateVM_Task, 'duration_secs': 0.332208} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.715016] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1244.715823] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9c559b8-09bf-41ff-8b1b-16f309dd363f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1244.716018] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9c559b8-09bf-41ff-8b1b-16f309dd363f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1244.716401] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a9c559b8-09bf-41ff-8b1b-16f309dd363f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1244.716681] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88a5b081-6115-4bac-b497-86dcabd5feea {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.723023] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1244.723023] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52bbd5dd-520a-8249-5bdf-fdce45b9ac39" [ 1244.723023] env[62692]: _type = "Task" [ 1244.723023] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.733261] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52bbd5dd-520a-8249-5bdf-fdce45b9ac39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.232260] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9c559b8-09bf-41ff-8b1b-16f309dd363f" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1245.232572] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Processing image a9c559b8-09bf-41ff-8b1b-16f309dd363f {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1245.232810] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a9c559b8-09bf-41ff-8b1b-16f309dd363f/a9c559b8-09bf-41ff-8b1b-16f309dd363f.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1245.232987] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a9c559b8-09bf-41ff-8b1b-16f309dd363f/a9c559b8-09bf-41ff-8b1b-16f309dd363f.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1245.233242] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1245.233503] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ca92d848-7ce1-47bf-9b34-9ae0e78c10d1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.252589] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1245.252589] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1245.253276] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c8e4da5-b7fc-464c-ab67-206d67e14045 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.259303] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1245.259303] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5218f0fe-c8ac-1f9b-b5e2-c15c183b46f8" [ 1245.259303] env[62692]: _type = "Task" [ 1245.259303] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.266937] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5218f0fe-c8ac-1f9b-b5e2-c15c183b46f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.273135] env[62692]: DEBUG nova.network.neutron [req-b03a9d8b-5ff6-4cb7-b7ed-3decb04b560f req-261ea049-861f-499d-b041-5e98db28ba49 service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Updated VIF entry in instance network info cache for port d463b030-eded-4099-ae6f-2aea2d1af078. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1245.273513] env[62692]: DEBUG nova.network.neutron [req-b03a9d8b-5ff6-4cb7-b7ed-3decb04b560f req-261ea049-861f-499d-b041-5e98db28ba49 service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Updating instance_info_cache with network_info: [{"id": "d463b030-eded-4099-ae6f-2aea2d1af078", "address": "fa:16:3e:db:5f:6a", "network": {"id": "189f2464-2a0a-4637-b0f3-2f6ff8fdb324", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1462324860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f94f7500bf524e3890734a58937b80c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd463b030-ed", "ovs_interfaceid": "d463b030-eded-4099-ae6f-2aea2d1af078", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1245.769936] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Preparing fetch location {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1245.770261] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Fetch image to [datastore2] OSTACK_IMG_937c167d-0edf-48f3-96fb-63d0e883704b/OSTACK_IMG_937c167d-0edf-48f3-96fb-63d0e883704b.vmdk {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1245.770455] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Downloading stream optimized image a9c559b8-09bf-41ff-8b1b-16f309dd363f to [datastore2] OSTACK_IMG_937c167d-0edf-48f3-96fb-63d0e883704b/OSTACK_IMG_937c167d-0edf-48f3-96fb-63d0e883704b.vmdk on the data store datastore2 as vApp {{(pid=62692) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1245.770581] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Downloading image file data a9c559b8-09bf-41ff-8b1b-16f309dd363f to the ESX as VM named 'OSTACK_IMG_937c167d-0edf-48f3-96fb-63d0e883704b' {{(pid=62692) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1245.777452] env[62692]: DEBUG oslo_concurrency.lockutils [req-b03a9d8b-5ff6-4cb7-b7ed-3decb04b560f req-261ea049-861f-499d-b041-5e98db28ba49 service nova] Releasing lock "refresh_cache-8abacc6f-054f-4c4f-abb2-83f20748cf1c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1245.854717] env[62692]: DEBUG oslo_vmware.rw_handles [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1245.854717] env[62692]: value = "resgroup-9" [ 1245.854717] env[62692]: _type = "ResourcePool" [ 1245.854717] env[62692]: }. {{(pid=62692) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1245.855040] env[62692]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-755ebf8e-c908-42e3-86dc-7c354402ac3f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.877398] env[62692]: DEBUG oslo_vmware.rw_handles [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lease: (returnval){ [ 1245.877398] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52535eae-71f5-2795-96d7-43a8fe27586f" [ 1245.877398] env[62692]: _type = "HttpNfcLease" [ 1245.877398] env[62692]: } obtained for vApp import into resource pool (val){ [ 1245.877398] env[62692]: value = "resgroup-9" [ 1245.877398] env[62692]: _type = "ResourcePool" [ 1245.877398] env[62692]: }. {{(pid=62692) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1245.877794] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the lease: (returnval){ [ 1245.877794] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52535eae-71f5-2795-96d7-43a8fe27586f" [ 1245.877794] env[62692]: _type = "HttpNfcLease" [ 1245.877794] env[62692]: } to be ready. {{(pid=62692) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1245.884112] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1245.884112] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52535eae-71f5-2795-96d7-43a8fe27586f" [ 1245.884112] env[62692]: _type = "HttpNfcLease" [ 1245.884112] env[62692]: } is initializing. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1246.389038] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1246.389038] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52535eae-71f5-2795-96d7-43a8fe27586f" [ 1246.389038] env[62692]: _type = "HttpNfcLease" [ 1246.389038] env[62692]: } is ready. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1246.389038] env[62692]: DEBUG oslo_vmware.rw_handles [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1246.389038] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52535eae-71f5-2795-96d7-43a8fe27586f" [ 1246.389038] env[62692]: _type = "HttpNfcLease" [ 1246.389038] env[62692]: }. {{(pid=62692) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1246.389830] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3de6a624-585f-4a1b-b6de-d6e9b8d2243a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.397875] env[62692]: DEBUG oslo_vmware.rw_handles [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52bf5bb0-abc1-f4f3-397c-665260e609ca/disk-0.vmdk from lease info. {{(pid=62692) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1246.398126] env[62692]: DEBUG oslo_vmware.rw_handles [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Creating HTTP connection to write to file with size = 31666688 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52bf5bb0-abc1-f4f3-397c-665260e609ca/disk-0.vmdk. {{(pid=62692) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1246.462341] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-efeb1a04-45f9-4a35-a85c-9e45373fa814 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.491798] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1247.491980] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1247.492149] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Starting heal instance info cache {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1247.682050] env[62692]: DEBUG oslo_vmware.rw_handles [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Completed reading data from the image iterator. {{(pid=62692) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1247.682050] env[62692]: DEBUG oslo_vmware.rw_handles [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52bf5bb0-abc1-f4f3-397c-665260e609ca/disk-0.vmdk. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1247.682946] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1056abe-ddde-4140-8931-2fb2d352f5a1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.690635] env[62692]: DEBUG oslo_vmware.rw_handles [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52bf5bb0-abc1-f4f3-397c-665260e609ca/disk-0.vmdk is in state: ready. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1247.690870] env[62692]: DEBUG oslo_vmware.rw_handles [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52bf5bb0-abc1-f4f3-397c-665260e609ca/disk-0.vmdk. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1247.691215] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-74657471-8107-4774-83d1-f1b4128ec64c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.000352] env[62692]: DEBUG oslo_vmware.rw_handles [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52bf5bb0-abc1-f4f3-397c-665260e609ca/disk-0.vmdk. {{(pid=62692) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1248.000604] env[62692]: INFO nova.virt.vmwareapi.images [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Downloaded image file data a9c559b8-09bf-41ff-8b1b-16f309dd363f [ 1248.001531] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18975d38-d8f6-4a19-b4cc-05f5e2f65cc0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.019579] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cfb27b08-4743-43d3-a2a0-ba33050c5835 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.064338] env[62692]: INFO nova.virt.vmwareapi.images [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] The imported VM was unregistered [ 1248.067467] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Caching image {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1248.067467] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Creating directory with path [datastore2] devstack-image-cache_base/a9c559b8-09bf-41ff-8b1b-16f309dd363f {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1248.067645] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-12b2d1f9-2a5d-4e9a-8ce2-19c1cd69d4fd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.089779] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Created directory with path [datastore2] devstack-image-cache_base/a9c559b8-09bf-41ff-8b1b-16f309dd363f {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1248.090011] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_937c167d-0edf-48f3-96fb-63d0e883704b/OSTACK_IMG_937c167d-0edf-48f3-96fb-63d0e883704b.vmdk to [datastore2] devstack-image-cache_base/a9c559b8-09bf-41ff-8b1b-16f309dd363f/a9c559b8-09bf-41ff-8b1b-16f309dd363f.vmdk. {{(pid=62692) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1248.090300] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-e602887b-954b-4940-b3da-40c083a9f7fa {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.097208] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1248.097208] env[62692]: value = "task-1142167" [ 1248.097208] env[62692]: _type = "Task" [ 1248.097208] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1248.107054] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142167, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.611116] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142167, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1249.111884] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142167, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1249.613201] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142167, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.037091] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "refresh_cache-c2f2a485-363e-41af-9b3f-37092947ca08" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1250.037262] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquired lock "refresh_cache-c2f2a485-363e-41af-9b3f-37092947ca08" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1250.037439] env[62692]: DEBUG nova.network.neutron [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Forcefully refreshing network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1250.113314] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142167, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.610378] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142167, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.259473} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.610915] env[62692]: INFO nova.virt.vmwareapi.ds_util [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_937c167d-0edf-48f3-96fb-63d0e883704b/OSTACK_IMG_937c167d-0edf-48f3-96fb-63d0e883704b.vmdk to [datastore2] devstack-image-cache_base/a9c559b8-09bf-41ff-8b1b-16f309dd363f/a9c559b8-09bf-41ff-8b1b-16f309dd363f.vmdk. [ 1250.611240] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Cleaning up location [datastore2] OSTACK_IMG_937c167d-0edf-48f3-96fb-63d0e883704b {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1250.611431] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_937c167d-0edf-48f3-96fb-63d0e883704b {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1250.611678] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-03a47609-430b-4c92-ae68-e7f0e26fced1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.617494] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1250.617494] env[62692]: value = "task-1142168" [ 1250.617494] env[62692]: _type = "Task" [ 1250.617494] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.625050] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142168, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.127596] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142168, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.061609} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.127858] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1251.128041] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a9c559b8-09bf-41ff-8b1b-16f309dd363f/a9c559b8-09bf-41ff-8b1b-16f309dd363f.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1251.128294] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a9c559b8-09bf-41ff-8b1b-16f309dd363f/a9c559b8-09bf-41ff-8b1b-16f309dd363f.vmdk to [datastore2] 8abacc6f-054f-4c4f-abb2-83f20748cf1c/8abacc6f-054f-4c4f-abb2-83f20748cf1c.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1251.128547] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8c959b31-e1ec-4ed6-8cbd-6bf72e7cdaad {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.134841] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1251.134841] env[62692]: value = "task-1142169" [ 1251.134841] env[62692]: _type = "Task" [ 1251.134841] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1251.142590] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142169, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.251552] env[62692]: DEBUG nova.network.neutron [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Updating instance_info_cache with network_info: [{"id": "fa8d79b4-2bf8-4b6b-a5c2-6893f618f420", "address": "fa:16:3e:a0:b7:ee", "network": {"id": "22795f7f-18d7-4b10-8051-bcff4e0af3ac", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-445901728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.250", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96b921004f8a4bf68640c3810fe7b819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa8d79b4-2b", "ovs_interfaceid": "fa8d79b4-2bf8-4b6b-a5c2-6893f618f420", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1251.648985] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142169, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.754925] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Releasing lock "refresh_cache-c2f2a485-363e-41af-9b3f-37092947ca08" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1251.755307] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Updated the network info_cache for instance {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1251.755451] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1251.755593] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1251.755750] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1251.755889] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1251.756050] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1251.756201] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._sync_power_states {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1252.152211] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142169, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.260117] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Getting list of instances from cluster (obj){ [ 1252.260117] env[62692]: value = "domain-c8" [ 1252.260117] env[62692]: _type = "ClusterComputeResource" [ 1252.260117] env[62692]: } {{(pid=62692) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 1252.261278] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4cf7f54-3636-4373-bdf4-18345b056dd6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.278420] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Got total of 3 instances {{(pid=62692) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 1252.278615] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Triggering sync for uuid c2f2a485-363e-41af-9b3f-37092947ca08 {{(pid=62692) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1252.278838] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Triggering sync for uuid 8abacc6f-054f-4c4f-abb2-83f20748cf1c {{(pid=62692) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1252.278956] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Triggering sync for uuid cb18c2e7-3241-4c9a-b1eb-a147b40530a1 {{(pid=62692) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1252.279331] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "c2f2a485-363e-41af-9b3f-37092947ca08" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1252.279671] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "c2f2a485-363e-41af-9b3f-37092947ca08" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1252.279858] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "8abacc6f-054f-4c4f-abb2-83f20748cf1c" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1252.280091] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "cb18c2e7-3241-4c9a-b1eb-a147b40530a1" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1252.280285] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "cb18c2e7-3241-4c9a-b1eb-a147b40530a1" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1252.280499] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1252.280637] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62692) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1252.281458] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c151a103-cfc9-4955-be26-247e466f151f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.284806] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15d3aeb3-e64c-45a7-b946-59f0b9934e7a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.287432] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager.update_available_resource {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1252.649638] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142169, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.790464] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1252.790768] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1252.790918] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1252.791152] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62692) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1252.792116] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96825fc0-6221-48fa-b2dc-ffed5991ba9a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.801089] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "cb18c2e7-3241-4c9a-b1eb-a147b40530a1" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.521s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1252.801518] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "c2f2a485-363e-41af-9b3f-37092947ca08" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.522s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1252.805294] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c5fdee6-a3e0-4d97-aa43-5b5a821212aa {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.823474] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a583d1b-277e-4958-bc95-e51a1ddb6d21 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.835236] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a12524b-5d85-446d-aac9-ffb8c62d415f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.868408] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180053MB free_disk=166GB free_vcpus=48 pci_devices=None {{(pid=62692) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1252.868622] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1252.868799] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1253.152671] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142169, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1253.647481] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142169, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.286752} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1253.647737] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a9c559b8-09bf-41ff-8b1b-16f309dd363f/a9c559b8-09bf-41ff-8b1b-16f309dd363f.vmdk to [datastore2] 8abacc6f-054f-4c4f-abb2-83f20748cf1c/8abacc6f-054f-4c4f-abb2-83f20748cf1c.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1253.648516] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b9fdce3-ad1d-4244-a389-273f64fca2ee {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.669860] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] 8abacc6f-054f-4c4f-abb2-83f20748cf1c/8abacc6f-054f-4c4f-abb2-83f20748cf1c.vmdk or device None with type streamOptimized {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1253.670254] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5055902e-5d5e-4271-a8a6-013a8b5eea9d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.688844] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1253.688844] env[62692]: value = "task-1142170" [ 1253.688844] env[62692]: _type = "Task" [ 1253.688844] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.696241] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142170, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1253.896256] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance c2f2a485-363e-41af-9b3f-37092947ca08 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1253.896428] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance cb18c2e7-3241-4c9a-b1eb-a147b40530a1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1253.896559] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 8abacc6f-054f-4c4f-abb2-83f20748cf1c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1253.896745] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62692) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1253.896885] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62692) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1253.945029] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c6c10ce-3af1-4315-995c-dae17a332b41 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.952707] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab3d1417-9e6c-4642-94ff-abf66c156e44 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.981568] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d09a2b99-dadf-4aec-8a02-3152a7ebf0af {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.988466] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c101b6e-44e2-47ac-80fc-0ff412446df2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.003979] env[62692]: DEBUG nova.compute.provider_tree [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1254.198914] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142170, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.506922] env[62692]: DEBUG nova.scheduler.client.report [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1254.700070] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142170, 'name': ReconfigVM_Task, 'duration_secs': 0.548281} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1254.700495] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Reconfigured VM instance instance-00000071 to attach disk [datastore2] 8abacc6f-054f-4c4f-abb2-83f20748cf1c/8abacc6f-054f-4c4f-abb2-83f20748cf1c.vmdk or device None with type streamOptimized {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1254.701097] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6ea3ee24-3ae5-4dd0-8362-23d46e4a111f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.707449] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1254.707449] env[62692]: value = "task-1142171" [ 1254.707449] env[62692]: _type = "Task" [ 1254.707449] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1254.715706] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142171, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.011976] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62692) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1255.012220] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.143s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1255.217434] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142171, 'name': Rename_Task, 'duration_secs': 0.139814} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1255.217720] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1255.217954] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1c1f679a-56b6-4ab8-9856-fdb601fb2a20 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.223936] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1255.223936] env[62692]: value = "task-1142172" [ 1255.223936] env[62692]: _type = "Task" [ 1255.223936] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.231299] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142172, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.733910] env[62692]: DEBUG oslo_vmware.api [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142172, 'name': PowerOnVM_Task, 'duration_secs': 0.424858} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1255.734337] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1255.869831] env[62692]: DEBUG nova.compute.manager [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1255.870768] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b20d005-d5a9-4ec5-99b5-d14e4d251f38 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.386423] env[62692]: DEBUG oslo_concurrency.lockutils [None req-1d979324-8d18-48eb-bf24-3f73d6ae4cb7 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "8abacc6f-054f-4c4f-abb2-83f20748cf1c" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 19.352s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1256.387413] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "8abacc6f-054f-4c4f-abb2-83f20748cf1c" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 4.107s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1256.387413] env[62692]: INFO nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] During sync_power_state the instance has a pending task (spawning). Skip. [ 1256.387550] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "8abacc6f-054f-4c4f-abb2-83f20748cf1c" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1257.864738] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d5c7ebc-f67d-47cb-9575-9f0022200f51 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.871698] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-317fa964-47a0-41d1-9b9a-09c190654378 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Suspending the VM {{(pid=62692) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1257.871930] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-8c94b2bb-22d6-4a06-a139-53ee830d68d9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.878489] env[62692]: DEBUG oslo_vmware.api [None req-317fa964-47a0-41d1-9b9a-09c190654378 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1257.878489] env[62692]: value = "task-1142173" [ 1257.878489] env[62692]: _type = "Task" [ 1257.878489] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.886459] env[62692]: DEBUG oslo_vmware.api [None req-317fa964-47a0-41d1-9b9a-09c190654378 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142173, 'name': SuspendVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.388650] env[62692]: DEBUG oslo_vmware.api [None req-317fa964-47a0-41d1-9b9a-09c190654378 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142173, 'name': SuspendVM_Task} progress is 75%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.889960] env[62692]: DEBUG oslo_vmware.api [None req-317fa964-47a0-41d1-9b9a-09c190654378 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142173, 'name': SuspendVM_Task, 'duration_secs': 0.550914} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.890369] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-317fa964-47a0-41d1-9b9a-09c190654378 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Suspended the VM {{(pid=62692) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1258.890369] env[62692]: DEBUG nova.compute.manager [None req-317fa964-47a0-41d1-9b9a-09c190654378 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1258.891117] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb26fde-5b43-4a9f-986c-afb287175b63 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.851429] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e14d1e2f-788e-4aae-8ba3-1ac184c28bfc tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "c2f2a485-363e-41af-9b3f-37092947ca08" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1259.851723] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e14d1e2f-788e-4aae-8ba3-1ac184c28bfc tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "c2f2a485-363e-41af-9b3f-37092947ca08" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1260.265123] env[62692]: INFO nova.compute.manager [None req-a442b028-f089-4ff4-a7ad-0d9ebef1230a tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Resuming [ 1260.265983] env[62692]: DEBUG nova.objects.instance [None req-a442b028-f089-4ff4-a7ad-0d9ebef1230a tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lazy-loading 'flavor' on Instance uuid 8abacc6f-054f-4c4f-abb2-83f20748cf1c {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1260.354965] env[62692]: INFO nova.compute.manager [None req-e14d1e2f-788e-4aae-8ba3-1ac184c28bfc tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Detaching volume c77c2006-709b-42d0-a47c-106f2a3f7237 [ 1260.384048] env[62692]: INFO nova.virt.block_device [None req-e14d1e2f-788e-4aae-8ba3-1ac184c28bfc tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Attempting to driver detach volume c77c2006-709b-42d0-a47c-106f2a3f7237 from mountpoint /dev/sdb [ 1260.384341] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-e14d1e2f-788e-4aae-8ba3-1ac184c28bfc tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Volume detach. Driver type: vmdk {{(pid=62692) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1260.384556] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-e14d1e2f-788e-4aae-8ba3-1ac184c28bfc tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249089', 'volume_id': 'c77c2006-709b-42d0-a47c-106f2a3f7237', 'name': 'volume-c77c2006-709b-42d0-a47c-106f2a3f7237', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'c2f2a485-363e-41af-9b3f-37092947ca08', 'attached_at': '', 'detached_at': '', 'volume_id': 'c77c2006-709b-42d0-a47c-106f2a3f7237', 'serial': 'c77c2006-709b-42d0-a47c-106f2a3f7237'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1260.385422] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a6292ef-0d53-41e3-b387-f2a9c99e042c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.406262] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a2ce3a4-a0ba-47ff-a5e0-61b067743c62 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.412616] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d09d472-4802-4476-9e6d-5c3fc8bf62c4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.433324] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d36ace-d44a-4c29-a8c0-0eaa506c3523 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.447208] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-e14d1e2f-788e-4aae-8ba3-1ac184c28bfc tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] The volume has not been displaced from its original location: [datastore2] volume-c77c2006-709b-42d0-a47c-106f2a3f7237/volume-c77c2006-709b-42d0-a47c-106f2a3f7237.vmdk. No consolidation needed. {{(pid=62692) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1260.452369] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-e14d1e2f-788e-4aae-8ba3-1ac184c28bfc tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Reconfiguring VM instance instance-0000006a to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1260.452630] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-66651ea8-1323-4d52-ac27-0121d0920e8a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.469983] env[62692]: DEBUG oslo_vmware.api [None req-e14d1e2f-788e-4aae-8ba3-1ac184c28bfc tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1260.469983] env[62692]: value = "task-1142174" [ 1260.469983] env[62692]: _type = "Task" [ 1260.469983] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.477221] env[62692]: DEBUG oslo_vmware.api [None req-e14d1e2f-788e-4aae-8ba3-1ac184c28bfc tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142174, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.980693] env[62692]: DEBUG oslo_vmware.api [None req-e14d1e2f-788e-4aae-8ba3-1ac184c28bfc tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142174, 'name': ReconfigVM_Task, 'duration_secs': 0.223573} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.980991] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-e14d1e2f-788e-4aae-8ba3-1ac184c28bfc tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Reconfigured VM instance instance-0000006a to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1260.985727] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-271241eb-7af0-419d-b2f6-bed6442617df {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.000471] env[62692]: DEBUG oslo_vmware.api [None req-e14d1e2f-788e-4aae-8ba3-1ac184c28bfc tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1261.000471] env[62692]: value = "task-1142175" [ 1261.000471] env[62692]: _type = "Task" [ 1261.000471] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1261.008083] env[62692]: DEBUG oslo_vmware.api [None req-e14d1e2f-788e-4aae-8ba3-1ac184c28bfc tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142175, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.273575] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a442b028-f089-4ff4-a7ad-0d9ebef1230a tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquiring lock "refresh_cache-8abacc6f-054f-4c4f-abb2-83f20748cf1c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1261.274035] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a442b028-f089-4ff4-a7ad-0d9ebef1230a tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquired lock "refresh_cache-8abacc6f-054f-4c4f-abb2-83f20748cf1c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1261.274035] env[62692]: DEBUG nova.network.neutron [None req-a442b028-f089-4ff4-a7ad-0d9ebef1230a tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1261.510764] env[62692]: DEBUG oslo_vmware.api [None req-e14d1e2f-788e-4aae-8ba3-1ac184c28bfc tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142175, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.965829] env[62692]: DEBUG nova.network.neutron [None req-a442b028-f089-4ff4-a7ad-0d9ebef1230a tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Updating instance_info_cache with network_info: [{"id": "d463b030-eded-4099-ae6f-2aea2d1af078", "address": "fa:16:3e:db:5f:6a", "network": {"id": "189f2464-2a0a-4637-b0f3-2f6ff8fdb324", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1462324860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f94f7500bf524e3890734a58937b80c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd463b030-ed", "ovs_interfaceid": "d463b030-eded-4099-ae6f-2aea2d1af078", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1262.011069] env[62692]: DEBUG oslo_vmware.api [None req-e14d1e2f-788e-4aae-8ba3-1ac184c28bfc tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142175, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.468356] env[62692]: DEBUG oslo_concurrency.lockutils [None req-a442b028-f089-4ff4-a7ad-0d9ebef1230a tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Releasing lock "refresh_cache-8abacc6f-054f-4c4f-abb2-83f20748cf1c" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1262.469350] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8504e354-fb72-48b5-b259-77e73f51d513 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.476044] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a442b028-f089-4ff4-a7ad-0d9ebef1230a tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Resuming the VM {{(pid=62692) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 1262.476292] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4233582c-9af6-4abf-a25a-5e839f10662f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.482712] env[62692]: DEBUG oslo_vmware.api [None req-a442b028-f089-4ff4-a7ad-0d9ebef1230a tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1262.482712] env[62692]: value = "task-1142176" [ 1262.482712] env[62692]: _type = "Task" [ 1262.482712] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.490364] env[62692]: DEBUG oslo_vmware.api [None req-a442b028-f089-4ff4-a7ad-0d9ebef1230a tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142176, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.510724] env[62692]: DEBUG oslo_vmware.api [None req-e14d1e2f-788e-4aae-8ba3-1ac184c28bfc tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142175, 'name': ReconfigVM_Task, 'duration_secs': 1.135225} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.511067] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-e14d1e2f-788e-4aae-8ba3-1ac184c28bfc tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249089', 'volume_id': 'c77c2006-709b-42d0-a47c-106f2a3f7237', 'name': 'volume-c77c2006-709b-42d0-a47c-106f2a3f7237', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'c2f2a485-363e-41af-9b3f-37092947ca08', 'attached_at': '', 'detached_at': '', 'volume_id': 'c77c2006-709b-42d0-a47c-106f2a3f7237', 'serial': 'c77c2006-709b-42d0-a47c-106f2a3f7237'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1262.993833] env[62692]: DEBUG oslo_vmware.api [None req-a442b028-f089-4ff4-a7ad-0d9ebef1230a tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142176, 'name': PowerOnVM_Task, 'duration_secs': 0.504225} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.994221] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-a442b028-f089-4ff4-a7ad-0d9ebef1230a tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Resumed the VM {{(pid=62692) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 1262.994510] env[62692]: DEBUG nova.compute.manager [None req-a442b028-f089-4ff4-a7ad-0d9ebef1230a tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1262.995489] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82f73096-6240-429f-887b-8555b96d71bb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.058133] env[62692]: DEBUG nova.objects.instance [None req-e14d1e2f-788e-4aae-8ba3-1ac184c28bfc tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lazy-loading 'flavor' on Instance uuid c2f2a485-363e-41af-9b3f-37092947ca08 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1264.067677] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e14d1e2f-788e-4aae-8ba3-1ac184c28bfc tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "c2f2a485-363e-41af-9b3f-37092947ca08" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.216s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1265.128725] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "c2f2a485-363e-41af-9b3f-37092947ca08" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1265.129106] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "c2f2a485-363e-41af-9b3f-37092947ca08" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1265.129434] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "c2f2a485-363e-41af-9b3f-37092947ca08-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1265.129694] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "c2f2a485-363e-41af-9b3f-37092947ca08-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1265.129928] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "c2f2a485-363e-41af-9b3f-37092947ca08-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1265.132572] env[62692]: INFO nova.compute.manager [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Terminating instance [ 1265.134754] env[62692]: DEBUG nova.compute.manager [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1265.135031] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1265.136133] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92936b99-0794-4159-95d0-ac70370bacb3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.146219] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1265.146501] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a174c05c-1151-48ce-94e2-b1e7dc7700d6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.153504] env[62692]: DEBUG oslo_vmware.api [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1265.153504] env[62692]: value = "task-1142177" [ 1265.153504] env[62692]: _type = "Task" [ 1265.153504] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.162907] env[62692]: DEBUG oslo_vmware.api [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142177, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.662202] env[62692]: DEBUG oslo_vmware.api [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142177, 'name': PowerOffVM_Task, 'duration_secs': 0.230669} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.662519] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1265.662704] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1265.662948] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-493282cb-3d14-4160-87a2-9cb6287930d3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.725205] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1265.725430] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1265.725623] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Deleting the datastore file [datastore1] c2f2a485-363e-41af-9b3f-37092947ca08 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1265.725895] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f955cb7d-63d0-4e70-81fd-e2d5b6ec6757 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.732396] env[62692]: DEBUG oslo_vmware.api [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1265.732396] env[62692]: value = "task-1142179" [ 1265.732396] env[62692]: _type = "Task" [ 1265.732396] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.739834] env[62692]: DEBUG oslo_vmware.api [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142179, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.242878] env[62692]: DEBUG oslo_vmware.api [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142179, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146178} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1266.243303] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1266.243378] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1266.243550] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1266.243731] env[62692]: INFO nova.compute.manager [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1266.243975] env[62692]: DEBUG oslo.service.loopingcall [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1266.244204] env[62692]: DEBUG nova.compute.manager [-] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1266.244336] env[62692]: DEBUG nova.network.neutron [-] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1266.650958] env[62692]: DEBUG nova.compute.manager [req-1cc5e122-f76c-43b6-bf78-501952f136a9 req-44d549bf-585b-4b8b-b0e2-4736db4b9566 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Received event network-vif-deleted-fa8d79b4-2bf8-4b6b-a5c2-6893f618f420 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1266.651503] env[62692]: INFO nova.compute.manager [req-1cc5e122-f76c-43b6-bf78-501952f136a9 req-44d549bf-585b-4b8b-b0e2-4736db4b9566 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Neutron deleted interface fa8d79b4-2bf8-4b6b-a5c2-6893f618f420; detaching it from the instance and deleting it from the info cache [ 1266.651503] env[62692]: DEBUG nova.network.neutron [req-1cc5e122-f76c-43b6-bf78-501952f136a9 req-44d549bf-585b-4b8b-b0e2-4736db4b9566 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1267.129592] env[62692]: DEBUG nova.network.neutron [-] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1267.154279] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-20abc791-6330-457c-88f4-4ff71cfd5432 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.164994] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2baa7c59-6725-4365-8a06-4f7ad024bc91 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.189531] env[62692]: DEBUG nova.compute.manager [req-1cc5e122-f76c-43b6-bf78-501952f136a9 req-44d549bf-585b-4b8b-b0e2-4736db4b9566 service nova] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Detach interface failed, port_id=fa8d79b4-2bf8-4b6b-a5c2-6893f618f420, reason: Instance c2f2a485-363e-41af-9b3f-37092947ca08 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1267.498701] env[62692]: DEBUG oslo_concurrency.lockutils [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquiring lock "8abacc6f-054f-4c4f-abb2-83f20748cf1c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1267.499043] env[62692]: DEBUG oslo_concurrency.lockutils [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "8abacc6f-054f-4c4f-abb2-83f20748cf1c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1267.499240] env[62692]: DEBUG oslo_concurrency.lockutils [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquiring lock "8abacc6f-054f-4c4f-abb2-83f20748cf1c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1267.499437] env[62692]: DEBUG oslo_concurrency.lockutils [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "8abacc6f-054f-4c4f-abb2-83f20748cf1c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1267.499647] env[62692]: DEBUG oslo_concurrency.lockutils [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "8abacc6f-054f-4c4f-abb2-83f20748cf1c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1267.501930] env[62692]: INFO nova.compute.manager [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Terminating instance [ 1267.504307] env[62692]: DEBUG nova.compute.manager [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1267.504512] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1267.505367] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64566993-93f8-4751-b7c4-afca941cd35c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.513456] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1267.513688] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7ac9d6ef-8a41-4832-a655-178e7508b4cf {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.519328] env[62692]: DEBUG oslo_vmware.api [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1267.519328] env[62692]: value = "task-1142180" [ 1267.519328] env[62692]: _type = "Task" [ 1267.519328] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1267.526647] env[62692]: DEBUG oslo_vmware.api [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142180, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.632538] env[62692]: INFO nova.compute.manager [-] [instance: c2f2a485-363e-41af-9b3f-37092947ca08] Took 1.39 seconds to deallocate network for instance. [ 1268.031584] env[62692]: DEBUG oslo_vmware.api [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142180, 'name': PowerOffVM_Task, 'duration_secs': 0.17325} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.032036] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1268.032339] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1268.032696] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9f6caa45-1716-4c14-8330-69c5e3550377 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.094486] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1268.094486] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1268.094734] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Deleting the datastore file [datastore2] 8abacc6f-054f-4c4f-abb2-83f20748cf1c {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1268.094839] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d664e77f-f52c-492a-94c5-a24cda2c2e55 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.100930] env[62692]: DEBUG oslo_vmware.api [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for the task: (returnval){ [ 1268.100930] env[62692]: value = "task-1142182" [ 1268.100930] env[62692]: _type = "Task" [ 1268.100930] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1268.108035] env[62692]: DEBUG oslo_vmware.api [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142182, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.139171] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1268.139433] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1268.139683] env[62692]: DEBUG nova.objects.instance [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lazy-loading 'resources' on Instance uuid c2f2a485-363e-41af-9b3f-37092947ca08 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1268.610819] env[62692]: DEBUG oslo_vmware.api [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Task: {'id': task-1142182, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141793} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.611645] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1268.611747] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1268.612831] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1268.612831] env[62692]: INFO nova.compute.manager [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1268.612831] env[62692]: DEBUG oslo.service.loopingcall [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1268.612831] env[62692]: DEBUG nova.compute.manager [-] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1268.612831] env[62692]: DEBUG nova.network.neutron [-] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1268.694884] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3793611f-040a-4742-a85b-40bc22bafde7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.702523] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb285a35-6918-491f-9b73-8860eb023f85 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.733635] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b73ef584-a9fa-4cb8-aa6e-dfa4d7d6f974 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.741039] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c2e85bb-08ed-4600-b57f-8ea74dd896fd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.754355] env[62692]: DEBUG nova.compute.provider_tree [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1268.878217] env[62692]: DEBUG nova.compute.manager [req-975f38c9-2c37-4f56-bbf5-2334b851710d req-ada1cb40-74f0-4929-adf1-deaa586fc7a6 service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Received event network-vif-deleted-d463b030-eded-4099-ae6f-2aea2d1af078 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1268.878471] env[62692]: INFO nova.compute.manager [req-975f38c9-2c37-4f56-bbf5-2334b851710d req-ada1cb40-74f0-4929-adf1-deaa586fc7a6 service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Neutron deleted interface d463b030-eded-4099-ae6f-2aea2d1af078; detaching it from the instance and deleting it from the info cache [ 1268.878697] env[62692]: DEBUG nova.network.neutron [req-975f38c9-2c37-4f56-bbf5-2334b851710d req-ada1cb40-74f0-4929-adf1-deaa586fc7a6 service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1269.257267] env[62692]: DEBUG nova.scheduler.client.report [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1269.349416] env[62692]: DEBUG nova.network.neutron [-] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1269.381315] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-04109b43-a45f-43ca-a635-e674ffe4f709 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.390375] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15ec0c4a-8d6c-4dd2-beee-962aa472c6b4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.414687] env[62692]: DEBUG nova.compute.manager [req-975f38c9-2c37-4f56-bbf5-2334b851710d req-ada1cb40-74f0-4929-adf1-deaa586fc7a6 service nova] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Detach interface failed, port_id=d463b030-eded-4099-ae6f-2aea2d1af078, reason: Instance 8abacc6f-054f-4c4f-abb2-83f20748cf1c could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1269.762952] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.623s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1269.785192] env[62692]: INFO nova.scheduler.client.report [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Deleted allocations for instance c2f2a485-363e-41af-9b3f-37092947ca08 [ 1269.852074] env[62692]: INFO nova.compute.manager [-] [instance: 8abacc6f-054f-4c4f-abb2-83f20748cf1c] Took 1.24 seconds to deallocate network for instance. [ 1270.293023] env[62692]: DEBUG oslo_concurrency.lockutils [None req-bd1ec42c-7370-4842-8802-b242581468e9 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "c2f2a485-363e-41af-9b3f-37092947ca08" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.164s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1270.358424] env[62692]: DEBUG oslo_concurrency.lockutils [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1270.358704] env[62692]: DEBUG oslo_concurrency.lockutils [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1270.358929] env[62692]: DEBUG nova.objects.instance [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lazy-loading 'resources' on Instance uuid 8abacc6f-054f-4c4f-abb2-83f20748cf1c {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1270.903169] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f5c6c39-2b68-422a-9a9e-939582c201f0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.910958] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc7ba678-e263-4c77-98ea-9e9a157829f5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.941110] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a6b8e41-8bb7-4ace-bb6f-fba1f5cf01d6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.948459] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee4b721f-7103-42f2-afc9-c54d920bbd7f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.961528] env[62692]: DEBUG nova.compute.provider_tree [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1271.465055] env[62692]: DEBUG nova.scheduler.client.report [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1271.969974] env[62692]: DEBUG oslo_concurrency.lockutils [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.611s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1271.994463] env[62692]: INFO nova.scheduler.client.report [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Deleted allocations for instance 8abacc6f-054f-4c4f-abb2-83f20748cf1c [ 1272.078828] env[62692]: DEBUG oslo_concurrency.lockutils [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "e4f1850a-5310-4ee7-a3b7-adbba64f5a83" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1272.078977] env[62692]: DEBUG oslo_concurrency.lockutils [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "e4f1850a-5310-4ee7-a3b7-adbba64f5a83" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1272.502676] env[62692]: DEBUG oslo_concurrency.lockutils [None req-271a0dfa-4eea-4be5-b788-37a748536030 tempest-ServersNegativeTestJSON-242413790 tempest-ServersNegativeTestJSON-242413790-project-member] Lock "8abacc6f-054f-4c4f-abb2-83f20748cf1c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.004s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1272.581790] env[62692]: DEBUG nova.compute.manager [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1273.102994] env[62692]: DEBUG oslo_concurrency.lockutils [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1273.103328] env[62692]: DEBUG oslo_concurrency.lockutils [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1273.104857] env[62692]: INFO nova.compute.claims [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1274.158938] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e717632-03c2-4157-bffc-f77fd7bc7a43 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.167174] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c608c799-ea77-442f-af37-c3f64e53e058 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.196489] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3485bec-c827-4dc0-abc9-427ddd56bd80 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.203248] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5571d80c-34a7-40d8-b887-27c4a630dccb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.217045] env[62692]: DEBUG nova.compute.provider_tree [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1274.719962] env[62692]: DEBUG nova.scheduler.client.report [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1275.225830] env[62692]: DEBUG oslo_concurrency.lockutils [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.122s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1275.226460] env[62692]: DEBUG nova.compute.manager [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1275.731932] env[62692]: DEBUG nova.compute.utils [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1275.733391] env[62692]: DEBUG nova.compute.manager [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1275.733580] env[62692]: DEBUG nova.network.neutron [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1275.782795] env[62692]: DEBUG nova.policy [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '66364879aaea4bd2a4880b720da868ed', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '96b921004f8a4bf68640c3810fe7b819', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1276.236926] env[62692]: DEBUG nova.network.neutron [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Successfully created port: 1863d5f5-d859-4e5b-9d7f-b6e0df952e3d {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1276.239514] env[62692]: DEBUG nova.compute.manager [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1277.248999] env[62692]: DEBUG nova.compute.manager [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1277.273982] env[62692]: DEBUG nova.virt.hardware [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1277.274255] env[62692]: DEBUG nova.virt.hardware [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1277.274435] env[62692]: DEBUG nova.virt.hardware [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1277.274639] env[62692]: DEBUG nova.virt.hardware [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1277.274797] env[62692]: DEBUG nova.virt.hardware [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1277.274952] env[62692]: DEBUG nova.virt.hardware [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1277.275181] env[62692]: DEBUG nova.virt.hardware [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1277.275348] env[62692]: DEBUG nova.virt.hardware [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1277.275527] env[62692]: DEBUG nova.virt.hardware [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1277.275696] env[62692]: DEBUG nova.virt.hardware [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1277.275876] env[62692]: DEBUG nova.virt.hardware [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1277.276773] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e603acc7-46f8-4805-8703-d4a0bb3b9645 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.284896] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a485565-8934-47d2-8d60-f3722c0afe0a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.604699] env[62692]: DEBUG nova.compute.manager [req-72aabc47-7bfe-4d53-9ab0-2627d9b2865b req-fc184bcf-1524-4879-8e32-429a4bfdfea2 service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Received event network-vif-plugged-1863d5f5-d859-4e5b-9d7f-b6e0df952e3d {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1277.605045] env[62692]: DEBUG oslo_concurrency.lockutils [req-72aabc47-7bfe-4d53-9ab0-2627d9b2865b req-fc184bcf-1524-4879-8e32-429a4bfdfea2 service nova] Acquiring lock "e4f1850a-5310-4ee7-a3b7-adbba64f5a83-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1277.605193] env[62692]: DEBUG oslo_concurrency.lockutils [req-72aabc47-7bfe-4d53-9ab0-2627d9b2865b req-fc184bcf-1524-4879-8e32-429a4bfdfea2 service nova] Lock "e4f1850a-5310-4ee7-a3b7-adbba64f5a83-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1277.605345] env[62692]: DEBUG oslo_concurrency.lockutils [req-72aabc47-7bfe-4d53-9ab0-2627d9b2865b req-fc184bcf-1524-4879-8e32-429a4bfdfea2 service nova] Lock "e4f1850a-5310-4ee7-a3b7-adbba64f5a83-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1277.606216] env[62692]: DEBUG nova.compute.manager [req-72aabc47-7bfe-4d53-9ab0-2627d9b2865b req-fc184bcf-1524-4879-8e32-429a4bfdfea2 service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] No waiting events found dispatching network-vif-plugged-1863d5f5-d859-4e5b-9d7f-b6e0df952e3d {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1277.606216] env[62692]: WARNING nova.compute.manager [req-72aabc47-7bfe-4d53-9ab0-2627d9b2865b req-fc184bcf-1524-4879-8e32-429a4bfdfea2 service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Received unexpected event network-vif-plugged-1863d5f5-d859-4e5b-9d7f-b6e0df952e3d for instance with vm_state building and task_state spawning. [ 1277.684410] env[62692]: DEBUG nova.network.neutron [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Successfully updated port: 1863d5f5-d859-4e5b-9d7f-b6e0df952e3d {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1278.188637] env[62692]: DEBUG oslo_concurrency.lockutils [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "refresh_cache-e4f1850a-5310-4ee7-a3b7-adbba64f5a83" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1278.188774] env[62692]: DEBUG oslo_concurrency.lockutils [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquired lock "refresh_cache-e4f1850a-5310-4ee7-a3b7-adbba64f5a83" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1278.188936] env[62692]: DEBUG nova.network.neutron [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1278.720220] env[62692]: DEBUG nova.network.neutron [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1278.854017] env[62692]: DEBUG nova.network.neutron [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Updating instance_info_cache with network_info: [{"id": "1863d5f5-d859-4e5b-9d7f-b6e0df952e3d", "address": "fa:16:3e:3d:46:a6", "network": {"id": "22795f7f-18d7-4b10-8051-bcff4e0af3ac", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-445901728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96b921004f8a4bf68640c3810fe7b819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1863d5f5-d8", "ovs_interfaceid": "1863d5f5-d859-4e5b-9d7f-b6e0df952e3d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1279.317618] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b93ad39c-8798-44b7-9925-968a968db188 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "cb18c2e7-3241-4c9a-b1eb-a147b40530a1" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1279.317870] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b93ad39c-8798-44b7-9925-968a968db188 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "cb18c2e7-3241-4c9a-b1eb-a147b40530a1" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1279.356623] env[62692]: DEBUG oslo_concurrency.lockutils [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Releasing lock "refresh_cache-e4f1850a-5310-4ee7-a3b7-adbba64f5a83" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1279.356902] env[62692]: DEBUG nova.compute.manager [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Instance network_info: |[{"id": "1863d5f5-d859-4e5b-9d7f-b6e0df952e3d", "address": "fa:16:3e:3d:46:a6", "network": {"id": "22795f7f-18d7-4b10-8051-bcff4e0af3ac", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-445901728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96b921004f8a4bf68640c3810fe7b819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1863d5f5-d8", "ovs_interfaceid": "1863d5f5-d859-4e5b-9d7f-b6e0df952e3d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1279.357350] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3d:46:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd1da5fc2-0280-4f76-ac97-20ea4bc7bb16', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1863d5f5-d859-4e5b-9d7f-b6e0df952e3d', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1279.365528] env[62692]: DEBUG oslo.service.loopingcall [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1279.366072] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1279.366311] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-28445c82-0a2f-4858-8792-7125ce761508 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.386247] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1279.386247] env[62692]: value = "task-1142183" [ 1279.386247] env[62692]: _type = "Task" [ 1279.386247] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1279.393485] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1142183, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.631065] env[62692]: DEBUG nova.compute.manager [req-548ea8e5-590d-48bd-8c5d-f586e068fca9 req-9b81bfff-6258-4f55-9956-453623bec8e6 service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Received event network-changed-1863d5f5-d859-4e5b-9d7f-b6e0df952e3d {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1279.631357] env[62692]: DEBUG nova.compute.manager [req-548ea8e5-590d-48bd-8c5d-f586e068fca9 req-9b81bfff-6258-4f55-9956-453623bec8e6 service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Refreshing instance network info cache due to event network-changed-1863d5f5-d859-4e5b-9d7f-b6e0df952e3d. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1279.631604] env[62692]: DEBUG oslo_concurrency.lockutils [req-548ea8e5-590d-48bd-8c5d-f586e068fca9 req-9b81bfff-6258-4f55-9956-453623bec8e6 service nova] Acquiring lock "refresh_cache-e4f1850a-5310-4ee7-a3b7-adbba64f5a83" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1279.631825] env[62692]: DEBUG oslo_concurrency.lockutils [req-548ea8e5-590d-48bd-8c5d-f586e068fca9 req-9b81bfff-6258-4f55-9956-453623bec8e6 service nova] Acquired lock "refresh_cache-e4f1850a-5310-4ee7-a3b7-adbba64f5a83" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1279.632086] env[62692]: DEBUG nova.network.neutron [req-548ea8e5-590d-48bd-8c5d-f586e068fca9 req-9b81bfff-6258-4f55-9956-453623bec8e6 service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Refreshing network info cache for port 1863d5f5-d859-4e5b-9d7f-b6e0df952e3d {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1279.821625] env[62692]: DEBUG nova.compute.utils [None req-b93ad39c-8798-44b7-9925-968a968db188 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1279.896110] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1142183, 'name': CreateVM_Task, 'duration_secs': 0.27704} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1279.896267] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1279.896903] env[62692]: DEBUG oslo_concurrency.lockutils [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1279.897095] env[62692]: DEBUG oslo_concurrency.lockutils [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1279.897433] env[62692]: DEBUG oslo_concurrency.lockutils [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1279.897921] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8bdc9d40-7484-4fe5-8fca-164edde54b26 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.902131] env[62692]: DEBUG oslo_vmware.api [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1279.902131] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]521fbb92-ef36-040f-8de7-2216fc211800" [ 1279.902131] env[62692]: _type = "Task" [ 1279.902131] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1279.909399] env[62692]: DEBUG oslo_vmware.api [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]521fbb92-ef36-040f-8de7-2216fc211800, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1280.324159] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b93ad39c-8798-44b7-9925-968a968db188 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "cb18c2e7-3241-4c9a-b1eb-a147b40530a1" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1280.363150] env[62692]: DEBUG nova.network.neutron [req-548ea8e5-590d-48bd-8c5d-f586e068fca9 req-9b81bfff-6258-4f55-9956-453623bec8e6 service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Updated VIF entry in instance network info cache for port 1863d5f5-d859-4e5b-9d7f-b6e0df952e3d. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1280.363548] env[62692]: DEBUG nova.network.neutron [req-548ea8e5-590d-48bd-8c5d-f586e068fca9 req-9b81bfff-6258-4f55-9956-453623bec8e6 service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Updating instance_info_cache with network_info: [{"id": "1863d5f5-d859-4e5b-9d7f-b6e0df952e3d", "address": "fa:16:3e:3d:46:a6", "network": {"id": "22795f7f-18d7-4b10-8051-bcff4e0af3ac", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-445901728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96b921004f8a4bf68640c3810fe7b819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1863d5f5-d8", "ovs_interfaceid": "1863d5f5-d859-4e5b-9d7f-b6e0df952e3d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1280.412882] env[62692]: DEBUG oslo_vmware.api [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]521fbb92-ef36-040f-8de7-2216fc211800, 'name': SearchDatastore_Task, 'duration_secs': 0.011392} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1280.413198] env[62692]: DEBUG oslo_concurrency.lockutils [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1280.413459] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1280.413701] env[62692]: DEBUG oslo_concurrency.lockutils [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1280.413851] env[62692]: DEBUG oslo_concurrency.lockutils [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1280.414047] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1280.414303] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-db70bfe3-6716-4626-a6a6-7adfb7f8ed02 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.421971] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1280.422163] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1280.422841] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba3219fd-aa70-476e-b7f6-19e4f01c9871 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.427660] env[62692]: DEBUG oslo_vmware.api [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1280.427660] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52fc2b8a-b072-807a-cd69-351c5b37705c" [ 1280.427660] env[62692]: _type = "Task" [ 1280.427660] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1280.434661] env[62692]: DEBUG oslo_vmware.api [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52fc2b8a-b072-807a-cd69-351c5b37705c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1280.866740] env[62692]: DEBUG oslo_concurrency.lockutils [req-548ea8e5-590d-48bd-8c5d-f586e068fca9 req-9b81bfff-6258-4f55-9956-453623bec8e6 service nova] Releasing lock "refresh_cache-e4f1850a-5310-4ee7-a3b7-adbba64f5a83" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1280.938092] env[62692]: DEBUG oslo_vmware.api [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52fc2b8a-b072-807a-cd69-351c5b37705c, 'name': SearchDatastore_Task, 'duration_secs': 0.007955} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1280.938848] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c1ba7f8-e7a5-490f-b48b-aa21534f7693 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.943637] env[62692]: DEBUG oslo_vmware.api [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1280.943637] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52922178-980b-518d-16cd-2a9310945d40" [ 1280.943637] env[62692]: _type = "Task" [ 1280.943637] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1280.950554] env[62692]: DEBUG oslo_vmware.api [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52922178-980b-518d-16cd-2a9310945d40, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1281.390176] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b93ad39c-8798-44b7-9925-968a968db188 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "cb18c2e7-3241-4c9a-b1eb-a147b40530a1" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1281.390544] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b93ad39c-8798-44b7-9925-968a968db188 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "cb18c2e7-3241-4c9a-b1eb-a147b40530a1" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1281.390835] env[62692]: INFO nova.compute.manager [None req-b93ad39c-8798-44b7-9925-968a968db188 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Attaching volume 77bf14fd-1e80-4e41-a6a9-f70f1356b51f to /dev/sdb [ 1281.422995] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0a2b8e8-0f27-4465-abe9-770d49c88f52 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.429840] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c93206-62ba-4b3b-8da9-f1fff2448633 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.443043] env[62692]: DEBUG nova.virt.block_device [None req-b93ad39c-8798-44b7-9925-968a968db188 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Updating existing volume attachment record: b82e1f8e-c7e3-4f1f-acf6-a5537d69193a {{(pid=62692) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1281.452823] env[62692]: DEBUG oslo_vmware.api [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52922178-980b-518d-16cd-2a9310945d40, 'name': SearchDatastore_Task, 'duration_secs': 0.008876} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1281.453078] env[62692]: DEBUG oslo_concurrency.lockutils [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1281.453409] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] e4f1850a-5310-4ee7-a3b7-adbba64f5a83/e4f1850a-5310-4ee7-a3b7-adbba64f5a83.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1281.453635] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c88b0049-3b6a-4f6e-a27b-808be42313ea {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.459222] env[62692]: DEBUG oslo_vmware.api [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1281.459222] env[62692]: value = "task-1142184" [ 1281.459222] env[62692]: _type = "Task" [ 1281.459222] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1281.466421] env[62692]: DEBUG oslo_vmware.api [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142184, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1281.968965] env[62692]: DEBUG oslo_vmware.api [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142184, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.444028} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1281.969322] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] e4f1850a-5310-4ee7-a3b7-adbba64f5a83/e4f1850a-5310-4ee7-a3b7-adbba64f5a83.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1281.969548] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1281.969844] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aa1df8d5-0973-4bfe-92c0-277e69f8ac16 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.975951] env[62692]: DEBUG oslo_vmware.api [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1281.975951] env[62692]: value = "task-1142188" [ 1281.975951] env[62692]: _type = "Task" [ 1281.975951] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1281.983360] env[62692]: DEBUG oslo_vmware.api [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142188, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1282.485458] env[62692]: DEBUG oslo_vmware.api [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142188, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060291} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1282.485825] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1282.486656] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-186664be-368e-4b90-8b52-7afac264fab2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.508112] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Reconfiguring VM instance instance-00000079 to attach disk [datastore2] e4f1850a-5310-4ee7-a3b7-adbba64f5a83/e4f1850a-5310-4ee7-a3b7-adbba64f5a83.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1282.508357] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ab4690d-bdc6-4460-a1ab-80eddf6939e3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.527738] env[62692]: DEBUG oslo_vmware.api [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1282.527738] env[62692]: value = "task-1142189" [ 1282.527738] env[62692]: _type = "Task" [ 1282.527738] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1282.535052] env[62692]: DEBUG oslo_vmware.api [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142189, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1283.039115] env[62692]: DEBUG oslo_vmware.api [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142189, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1283.538013] env[62692]: DEBUG oslo_vmware.api [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142189, 'name': ReconfigVM_Task, 'duration_secs': 0.576728} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1283.538319] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Reconfigured VM instance instance-00000079 to attach disk [datastore2] e4f1850a-5310-4ee7-a3b7-adbba64f5a83/e4f1850a-5310-4ee7-a3b7-adbba64f5a83.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1283.538948] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-10679a34-1fa3-4e91-9fa7-191dcb66ee44 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.545643] env[62692]: DEBUG oslo_vmware.api [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1283.545643] env[62692]: value = "task-1142190" [ 1283.545643] env[62692]: _type = "Task" [ 1283.545643] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1283.553295] env[62692]: DEBUG oslo_vmware.api [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142190, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1284.057331] env[62692]: DEBUG oslo_vmware.api [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142190, 'name': Rename_Task, 'duration_secs': 0.129402} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1284.057690] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1284.058056] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1b3f4065-e975-4628-a712-99910c51f48a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.064531] env[62692]: DEBUG oslo_vmware.api [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1284.064531] env[62692]: value = "task-1142192" [ 1284.064531] env[62692]: _type = "Task" [ 1284.064531] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1284.072347] env[62692]: DEBUG oslo_vmware.api [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142192, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1284.574187] env[62692]: DEBUG oslo_vmware.api [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142192, 'name': PowerOnVM_Task, 'duration_secs': 0.459687} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1284.574508] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1284.574794] env[62692]: INFO nova.compute.manager [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Took 7.33 seconds to spawn the instance on the hypervisor. [ 1284.575038] env[62692]: DEBUG nova.compute.manager [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1284.575897] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1608720f-b9a7-43ef-9020-9ce42b7d8f15 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.094756] env[62692]: INFO nova.compute.manager [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Took 12.01 seconds to build instance. [ 1285.247042] env[62692]: DEBUG nova.compute.manager [req-396f5d35-5b3e-4b33-b2d6-f34e80a8e67e req-d3866b2f-2406-4041-9334-b12f32a2ed93 service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Received event network-changed-1863d5f5-d859-4e5b-9d7f-b6e0df952e3d {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1285.247266] env[62692]: DEBUG nova.compute.manager [req-396f5d35-5b3e-4b33-b2d6-f34e80a8e67e req-d3866b2f-2406-4041-9334-b12f32a2ed93 service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Refreshing instance network info cache due to event network-changed-1863d5f5-d859-4e5b-9d7f-b6e0df952e3d. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1285.247488] env[62692]: DEBUG oslo_concurrency.lockutils [req-396f5d35-5b3e-4b33-b2d6-f34e80a8e67e req-d3866b2f-2406-4041-9334-b12f32a2ed93 service nova] Acquiring lock "refresh_cache-e4f1850a-5310-4ee7-a3b7-adbba64f5a83" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1285.247638] env[62692]: DEBUG oslo_concurrency.lockutils [req-396f5d35-5b3e-4b33-b2d6-f34e80a8e67e req-d3866b2f-2406-4041-9334-b12f32a2ed93 service nova] Acquired lock "refresh_cache-e4f1850a-5310-4ee7-a3b7-adbba64f5a83" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1285.247806] env[62692]: DEBUG nova.network.neutron [req-396f5d35-5b3e-4b33-b2d6-f34e80a8e67e req-d3866b2f-2406-4041-9334-b12f32a2ed93 service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Refreshing network info cache for port 1863d5f5-d859-4e5b-9d7f-b6e0df952e3d {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1285.596896] env[62692]: DEBUG oslo_concurrency.lockutils [None req-78e7d053-7253-4aee-ad50-ece5b729c0be tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "e4f1850a-5310-4ee7-a3b7-adbba64f5a83" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.518s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1285.955856] env[62692]: DEBUG nova.network.neutron [req-396f5d35-5b3e-4b33-b2d6-f34e80a8e67e req-d3866b2f-2406-4041-9334-b12f32a2ed93 service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Updated VIF entry in instance network info cache for port 1863d5f5-d859-4e5b-9d7f-b6e0df952e3d. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1285.956276] env[62692]: DEBUG nova.network.neutron [req-396f5d35-5b3e-4b33-b2d6-f34e80a8e67e req-d3866b2f-2406-4041-9334-b12f32a2ed93 service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Updating instance_info_cache with network_info: [{"id": "1863d5f5-d859-4e5b-9d7f-b6e0df952e3d", "address": "fa:16:3e:3d:46:a6", "network": {"id": "22795f7f-18d7-4b10-8051-bcff4e0af3ac", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-445901728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96b921004f8a4bf68640c3810fe7b819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1863d5f5-d8", "ovs_interfaceid": "1863d5f5-d859-4e5b-9d7f-b6e0df952e3d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1285.986625] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-b93ad39c-8798-44b7-9925-968a968db188 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Volume attach. Driver type: vmdk {{(pid=62692) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1285.986857] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-b93ad39c-8798-44b7-9925-968a968db188 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249104', 'volume_id': '77bf14fd-1e80-4e41-a6a9-f70f1356b51f', 'name': 'volume-77bf14fd-1e80-4e41-a6a9-f70f1356b51f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cb18c2e7-3241-4c9a-b1eb-a147b40530a1', 'attached_at': '', 'detached_at': '', 'volume_id': '77bf14fd-1e80-4e41-a6a9-f70f1356b51f', 'serial': '77bf14fd-1e80-4e41-a6a9-f70f1356b51f'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1285.987736] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eec636e-8fa7-4f77-bc33-353367db72d3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.004015] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88e91a97-76a1-4436-bb9d-1b48b7d734b5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.027683] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-b93ad39c-8798-44b7-9925-968a968db188 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Reconfiguring VM instance instance-00000078 to attach disk [datastore2] volume-77bf14fd-1e80-4e41-a6a9-f70f1356b51f/volume-77bf14fd-1e80-4e41-a6a9-f70f1356b51f.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1286.027941] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b296d945-47fe-408a-988a-5331ab8d96ed {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.046191] env[62692]: DEBUG oslo_vmware.api [None req-b93ad39c-8798-44b7-9925-968a968db188 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1286.046191] env[62692]: value = "task-1142193" [ 1286.046191] env[62692]: _type = "Task" [ 1286.046191] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1286.054699] env[62692]: DEBUG oslo_vmware.api [None req-b93ad39c-8798-44b7-9925-968a968db188 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142193, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.458912] env[62692]: DEBUG oslo_concurrency.lockutils [req-396f5d35-5b3e-4b33-b2d6-f34e80a8e67e req-d3866b2f-2406-4041-9334-b12f32a2ed93 service nova] Releasing lock "refresh_cache-e4f1850a-5310-4ee7-a3b7-adbba64f5a83" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1286.556456] env[62692]: DEBUG oslo_vmware.api [None req-b93ad39c-8798-44b7-9925-968a968db188 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142193, 'name': ReconfigVM_Task, 'duration_secs': 0.376742} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1286.556737] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-b93ad39c-8798-44b7-9925-968a968db188 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Reconfigured VM instance instance-00000078 to attach disk [datastore2] volume-77bf14fd-1e80-4e41-a6a9-f70f1356b51f/volume-77bf14fd-1e80-4e41-a6a9-f70f1356b51f.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1286.561405] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c442efd-1aeb-4cc5-9c8a-3366871a5558 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.576021] env[62692]: DEBUG oslo_vmware.api [None req-b93ad39c-8798-44b7-9925-968a968db188 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1286.576021] env[62692]: value = "task-1142194" [ 1286.576021] env[62692]: _type = "Task" [ 1286.576021] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1286.583129] env[62692]: DEBUG oslo_vmware.api [None req-b93ad39c-8798-44b7-9925-968a968db188 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142194, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.085898] env[62692]: DEBUG oslo_vmware.api [None req-b93ad39c-8798-44b7-9925-968a968db188 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142194, 'name': ReconfigVM_Task, 'duration_secs': 0.136967} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1287.086156] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-b93ad39c-8798-44b7-9925-968a968db188 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249104', 'volume_id': '77bf14fd-1e80-4e41-a6a9-f70f1356b51f', 'name': 'volume-77bf14fd-1e80-4e41-a6a9-f70f1356b51f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cb18c2e7-3241-4c9a-b1eb-a147b40530a1', 'attached_at': '', 'detached_at': '', 'volume_id': '77bf14fd-1e80-4e41-a6a9-f70f1356b51f', 'serial': '77bf14fd-1e80-4e41-a6a9-f70f1356b51f'} {{(pid=62692) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1288.124707] env[62692]: DEBUG nova.objects.instance [None req-b93ad39c-8798-44b7-9925-968a968db188 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lazy-loading 'flavor' on Instance uuid cb18c2e7-3241-4c9a-b1eb-a147b40530a1 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1288.630212] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b93ad39c-8798-44b7-9925-968a968db188 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "cb18c2e7-3241-4c9a-b1eb-a147b40530a1" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.240s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1289.506495] env[62692]: DEBUG nova.compute.manager [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Stashing vm_state: active {{(pid=62692) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1289.705758] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1289.706022] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1290.027335] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1290.027676] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1290.210857] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1290.211090] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Starting heal instance info cache {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1290.534361] env[62692]: INFO nova.compute.claims [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1291.041202] env[62692]: INFO nova.compute.resource_tracker [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Updating resource usage from migration 0e906315-01e5-438f-a75d-1e4934a6174d [ 1291.091468] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-630b49b8-a53c-46ee-9ce0-7aadfdb77a4b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.099802] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e135b0b9-2b16-43c5-a477-4f1bcb5bafad {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.132219] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb3882f4-4c40-4794-969a-eb5a4ca0b720 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.139503] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9c196c9-81cc-432b-a855-503df2e7d046 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.153333] env[62692]: DEBUG nova.compute.provider_tree [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1291.217162] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Didn't find any instances for network info cache update. {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1291.217409] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1291.217603] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1291.217789] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1291.217981] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1291.218181] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1291.218355] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1291.218516] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62692) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1291.218703] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager.update_available_resource {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1291.656485] env[62692]: DEBUG nova.scheduler.client.report [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1291.721010] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1292.161703] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.134s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1292.161986] env[62692]: INFO nova.compute.manager [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Migrating [ 1292.170243] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.449s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1292.170462] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1292.170642] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62692) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1292.174580] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a66694b-9bad-474c-a7b9-1feb647b6e3e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.189015] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82fa318b-7243-4ea8-9b50-1391b6c2f418 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.204586] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40f1b38b-5a1f-4f68-b6e8-803183af9ff4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.211192] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e866a2d-2545-4afe-b6b8-805eea7aa6cc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.240652] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181321MB free_disk=166GB free_vcpus=48 pci_devices=None {{(pid=62692) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1292.240812] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1292.240978] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1292.684166] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "refresh_cache-cb18c2e7-3241-4c9a-b1eb-a147b40530a1" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1292.684562] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquired lock "refresh_cache-cb18c2e7-3241-4c9a-b1eb-a147b40530a1" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1292.684562] env[62692]: DEBUG nova.network.neutron [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1293.247613] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Applying migration context for instance cb18c2e7-3241-4c9a-b1eb-a147b40530a1 as it has an incoming, in-progress migration 0e906315-01e5-438f-a75d-1e4934a6174d. Migration status is pre-migrating {{(pid=62692) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1293.248253] env[62692]: INFO nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Updating resource usage from migration 0e906315-01e5-438f-a75d-1e4934a6174d [ 1293.268052] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance e4f1850a-5310-4ee7-a3b7-adbba64f5a83 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1293.268052] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Migration 0e906315-01e5-438f-a75d-1e4934a6174d is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1293.268052] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance cb18c2e7-3241-4c9a-b1eb-a147b40530a1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1293.268254] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62692) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1293.268254] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1152MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62692) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1293.310913] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e087b25-59cf-4e81-8d0a-75a6134c3a44 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.320443] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24b56066-af61-4008-a629-1410adc8e196 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.351952] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32346617-6c8f-4e5d-92a8-476edafd2c2c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.359154] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d47271-2c09-44d8-8a54-da722fbc09bd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.372068] env[62692]: DEBUG nova.compute.provider_tree [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1293.419777] env[62692]: DEBUG nova.network.neutron [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Updating instance_info_cache with network_info: [{"id": "f84cd756-9d8a-4da6-bcd3-c377d3a0a875", "address": "fa:16:3e:88:fe:19", "network": {"id": "77695f27-3027-4a92-8b9b-0a90cdc53f62", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-337681783-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697b4281872f4547ad3d051e644b95ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf84cd756-9d", "ovs_interfaceid": "f84cd756-9d8a-4da6-bcd3-c377d3a0a875", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1293.875564] env[62692]: DEBUG nova.scheduler.client.report [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1293.923162] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Releasing lock "refresh_cache-cb18c2e7-3241-4c9a-b1eb-a147b40530a1" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1294.380109] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62692) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1294.380311] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.139s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1295.438737] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54169fb0-3a98-4eb5-a5ac-8e31d27ec02b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.461265] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Updating instance 'cb18c2e7-3241-4c9a-b1eb-a147b40530a1' progress to 0 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1295.967775] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1295.968114] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-95571061-9dc3-4801-9c4d-a67d28812db1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.975435] env[62692]: DEBUG oslo_vmware.api [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1295.975435] env[62692]: value = "task-1142195" [ 1295.975435] env[62692]: _type = "Task" [ 1295.975435] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1295.982737] env[62692]: DEBUG oslo_vmware.api [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142195, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1296.485718] env[62692]: DEBUG oslo_vmware.api [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142195, 'name': PowerOffVM_Task, 'duration_secs': 0.222824} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1296.486088] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1296.486195] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Updating instance 'cb18c2e7-3241-4c9a-b1eb-a147b40530a1' progress to 17 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1296.994031] env[62692]: DEBUG nova.virt.hardware [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1296.994031] env[62692]: DEBUG nova.virt.hardware [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1296.994031] env[62692]: DEBUG nova.virt.hardware [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1296.994031] env[62692]: DEBUG nova.virt.hardware [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1296.994031] env[62692]: DEBUG nova.virt.hardware [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1296.994031] env[62692]: DEBUG nova.virt.hardware [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1296.994860] env[62692]: DEBUG nova.virt.hardware [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1296.995256] env[62692]: DEBUG nova.virt.hardware [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1296.995608] env[62692]: DEBUG nova.virt.hardware [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1296.998012] env[62692]: DEBUG nova.virt.hardware [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1296.998012] env[62692]: DEBUG nova.virt.hardware [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1297.001417] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2af2b1d6-0d1b-4158-bc35-78b08d8e583f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.017982] env[62692]: DEBUG oslo_vmware.api [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1297.017982] env[62692]: value = "task-1142196" [ 1297.017982] env[62692]: _type = "Task" [ 1297.017982] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1297.026709] env[62692]: DEBUG oslo_vmware.api [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142196, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1297.528343] env[62692]: DEBUG oslo_vmware.api [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142196, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1298.029032] env[62692]: DEBUG oslo_vmware.api [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142196, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1298.529365] env[62692]: DEBUG oslo_vmware.api [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142196, 'name': ReconfigVM_Task, 'duration_secs': 1.17071} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1298.529758] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Updating instance 'cb18c2e7-3241-4c9a-b1eb-a147b40530a1' progress to 33 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1299.036896] env[62692]: DEBUG nova.virt.hardware [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1299.037166] env[62692]: DEBUG nova.virt.hardware [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1299.037333] env[62692]: DEBUG nova.virt.hardware [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1299.037525] env[62692]: DEBUG nova.virt.hardware [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1299.037743] env[62692]: DEBUG nova.virt.hardware [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1299.037950] env[62692]: DEBUG nova.virt.hardware [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1299.038183] env[62692]: DEBUG nova.virt.hardware [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1299.038351] env[62692]: DEBUG nova.virt.hardware [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1299.038523] env[62692]: DEBUG nova.virt.hardware [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1299.038695] env[62692]: DEBUG nova.virt.hardware [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1299.038875] env[62692]: DEBUG nova.virt.hardware [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1299.044196] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Reconfiguring VM instance instance-00000078 to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1299.044488] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4922c5da-f788-49b6-a8db-e4b1f00b138f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.063352] env[62692]: DEBUG oslo_vmware.api [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1299.063352] env[62692]: value = "task-1142197" [ 1299.063352] env[62692]: _type = "Task" [ 1299.063352] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1299.070744] env[62692]: DEBUG oslo_vmware.api [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142197, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1299.573408] env[62692]: DEBUG oslo_vmware.api [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142197, 'name': ReconfigVM_Task, 'duration_secs': 0.236682} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1299.573791] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Reconfigured VM instance instance-00000078 to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1299.574530] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bb9da5e-179d-4d2e-8718-c0b3bae9550e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.599329] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Reconfiguring VM instance instance-00000078 to attach disk [datastore1] cb18c2e7-3241-4c9a-b1eb-a147b40530a1/cb18c2e7-3241-4c9a-b1eb-a147b40530a1.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1299.599797] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0535aa6f-1e63-490e-8ca4-9b6b23296fa6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.617487] env[62692]: DEBUG oslo_vmware.api [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1299.617487] env[62692]: value = "task-1142198" [ 1299.617487] env[62692]: _type = "Task" [ 1299.617487] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1299.624770] env[62692]: DEBUG oslo_vmware.api [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142198, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1300.126954] env[62692]: DEBUG oslo_vmware.api [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142198, 'name': ReconfigVM_Task, 'duration_secs': 0.314415} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1300.127269] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Reconfigured VM instance instance-00000078 to attach disk [datastore1] cb18c2e7-3241-4c9a-b1eb-a147b40530a1/cb18c2e7-3241-4c9a-b1eb-a147b40530a1.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1300.127559] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Updating instance 'cb18c2e7-3241-4c9a-b1eb-a147b40530a1' progress to 50 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1300.635073] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83f314c7-a8ae-4edd-8a79-75aea9d15510 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.658729] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eabfba4-bb4b-4889-9789-8a4ec65d0bd3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.680912] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Updating instance 'cb18c2e7-3241-4c9a-b1eb-a147b40530a1' progress to 67 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1302.316027] env[62692]: DEBUG nova.network.neutron [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Port f84cd756-9d8a-4da6-bcd3-c377d3a0a875 binding to destination host cpu-1 is already ACTIVE {{(pid=62692) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1303.339593] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "cb18c2e7-3241-4c9a-b1eb-a147b40530a1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1303.339964] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "cb18c2e7-3241-4c9a-b1eb-a147b40530a1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1303.340090] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "cb18c2e7-3241-4c9a-b1eb-a147b40530a1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1304.373237] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "refresh_cache-cb18c2e7-3241-4c9a-b1eb-a147b40530a1" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1304.373513] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquired lock "refresh_cache-cb18c2e7-3241-4c9a-b1eb-a147b40530a1" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1304.373627] env[62692]: DEBUG nova.network.neutron [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1305.070543] env[62692]: DEBUG nova.network.neutron [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Updating instance_info_cache with network_info: [{"id": "f84cd756-9d8a-4da6-bcd3-c377d3a0a875", "address": "fa:16:3e:88:fe:19", "network": {"id": "77695f27-3027-4a92-8b9b-0a90cdc53f62", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-337681783-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697b4281872f4547ad3d051e644b95ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf84cd756-9d", "ovs_interfaceid": "f84cd756-9d8a-4da6-bcd3-c377d3a0a875", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1305.573159] env[62692]: DEBUG oslo_concurrency.lockutils [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Releasing lock "refresh_cache-cb18c2e7-3241-4c9a-b1eb-a147b40530a1" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1306.083225] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-628f49e8-6a35-4b7f-a904-f094400f5718 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.091163] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ece3b69-8aaa-420a-9aed-0c174c719de9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.198122] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11a07869-5b62-411d-bbdd-01e03251f978 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.221590] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a39ec39-51ed-4062-a7ef-fd7f7c91984a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.228377] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Updating instance 'cb18c2e7-3241-4c9a-b1eb-a147b40530a1' progress to 83 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1307.734484] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1307.734779] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cd2d83be-22b5-40c4-845f-4ce8f08cc028 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.742765] env[62692]: DEBUG oslo_vmware.api [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1307.742765] env[62692]: value = "task-1142199" [ 1307.742765] env[62692]: _type = "Task" [ 1307.742765] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1307.750499] env[62692]: DEBUG oslo_vmware.api [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142199, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1308.253048] env[62692]: DEBUG oslo_vmware.api [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142199, 'name': PowerOnVM_Task, 'duration_secs': 0.421391} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1308.253424] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1308.253524] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-c51548c8-92fa-4647-bd5f-6b5646350cb3 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Updating instance 'cb18c2e7-3241-4c9a-b1eb-a147b40530a1' progress to 100 {{(pid=62692) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1310.449977] env[62692]: DEBUG nova.network.neutron [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Port f84cd756-9d8a-4da6-bcd3-c377d3a0a875 binding to destination host cpu-1 is already ACTIVE {{(pid=62692) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1310.450288] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "refresh_cache-cb18c2e7-3241-4c9a-b1eb-a147b40530a1" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1310.450417] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquired lock "refresh_cache-cb18c2e7-3241-4c9a-b1eb-a147b40530a1" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1310.450584] env[62692]: DEBUG nova.network.neutron [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1311.158449] env[62692]: DEBUG nova.network.neutron [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Updating instance_info_cache with network_info: [{"id": "f84cd756-9d8a-4da6-bcd3-c377d3a0a875", "address": "fa:16:3e:88:fe:19", "network": {"id": "77695f27-3027-4a92-8b9b-0a90cdc53f62", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-337681783-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697b4281872f4547ad3d051e644b95ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf84cd756-9d", "ovs_interfaceid": "f84cd756-9d8a-4da6-bcd3-c377d3a0a875", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1311.661428] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Releasing lock "refresh_cache-cb18c2e7-3241-4c9a-b1eb-a147b40530a1" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1312.165511] env[62692]: DEBUG nova.compute.manager [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62692) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1313.264535] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1313.264902] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1313.768220] env[62692]: DEBUG nova.objects.instance [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lazy-loading 'migration_context' on Instance uuid cb18c2e7-3241-4c9a-b1eb-a147b40530a1 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1314.329890] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef598528-7d26-4d4c-a462-99fb67472838 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.337689] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b9f8e50-d17d-447e-bc2d-af82327b5cd6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.366631] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edde91ac-603d-4d41-a097-32f7f350004d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.373720] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-484d93e8-09bb-49c1-8111-043c731e9f95 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.386649] env[62692]: DEBUG nova.compute.provider_tree [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1314.892054] env[62692]: DEBUG nova.scheduler.client.report [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1315.904263] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.639s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1317.439638] env[62692]: INFO nova.compute.manager [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Swapping old allocation on dict_keys(['470758a8-5efa-4875-aad5-f512a727752f']) held by migration 0e906315-01e5-438f-a75d-1e4934a6174d for instance [ 1317.462369] env[62692]: DEBUG nova.scheduler.client.report [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Overwriting current allocation {'allocations': {'470758a8-5efa-4875-aad5-f512a727752f': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 168}}, 'project_id': '697b4281872f4547ad3d051e644b95ce', 'user_id': '6102acb4c19642d8a8190d828f5b5dfc', 'consumer_generation': 1} on consumer cb18c2e7-3241-4c9a-b1eb-a147b40530a1 {{(pid=62692) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1317.542497] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "refresh_cache-cb18c2e7-3241-4c9a-b1eb-a147b40530a1" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1317.542712] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquired lock "refresh_cache-cb18c2e7-3241-4c9a-b1eb-a147b40530a1" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1317.542898] env[62692]: DEBUG nova.network.neutron [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1318.247168] env[62692]: DEBUG nova.network.neutron [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Updating instance_info_cache with network_info: [{"id": "f84cd756-9d8a-4da6-bcd3-c377d3a0a875", "address": "fa:16:3e:88:fe:19", "network": {"id": "77695f27-3027-4a92-8b9b-0a90cdc53f62", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-337681783-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697b4281872f4547ad3d051e644b95ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf84cd756-9d", "ovs_interfaceid": "f84cd756-9d8a-4da6-bcd3-c377d3a0a875", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1318.750589] env[62692]: DEBUG oslo_concurrency.lockutils [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Releasing lock "refresh_cache-cb18c2e7-3241-4c9a-b1eb-a147b40530a1" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1318.751613] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d45ff6a1-fffb-4502-aed1-f03731b70ad9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.758934] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38f676e3-e21a-4026-b1a2-d3d94c9677e1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.842029] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1319.842437] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-80dd2f5b-d394-47bd-9707-7bf88920354d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.849466] env[62692]: DEBUG oslo_vmware.api [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1319.849466] env[62692]: value = "task-1142200" [ 1319.849466] env[62692]: _type = "Task" [ 1319.849466] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1319.857206] env[62692]: DEBUG oslo_vmware.api [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142200, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1320.359573] env[62692]: DEBUG oslo_vmware.api [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142200, 'name': PowerOffVM_Task, 'duration_secs': 0.1976} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1320.359843] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1320.360515] env[62692]: DEBUG nova.virt.hardware [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1320.360738] env[62692]: DEBUG nova.virt.hardware [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1320.360903] env[62692]: DEBUG nova.virt.hardware [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1320.361117] env[62692]: DEBUG nova.virt.hardware [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1320.361279] env[62692]: DEBUG nova.virt.hardware [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1320.361433] env[62692]: DEBUG nova.virt.hardware [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1320.361643] env[62692]: DEBUG nova.virt.hardware [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1320.361838] env[62692]: DEBUG nova.virt.hardware [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1320.362045] env[62692]: DEBUG nova.virt.hardware [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1320.362227] env[62692]: DEBUG nova.virt.hardware [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1320.362409] env[62692]: DEBUG nova.virt.hardware [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1320.367452] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7fa42e94-82aa-4275-8c40-d3c465c7a358 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.383125] env[62692]: DEBUG oslo_vmware.api [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1320.383125] env[62692]: value = "task-1142201" [ 1320.383125] env[62692]: _type = "Task" [ 1320.383125] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1320.390743] env[62692]: DEBUG oslo_vmware.api [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142201, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1320.893166] env[62692]: DEBUG oslo_vmware.api [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142201, 'name': ReconfigVM_Task, 'duration_secs': 0.160898} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1320.894079] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d0d04ed-aaab-4d8e-904b-d5462c6c3d21 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.917067] env[62692]: DEBUG nova.virt.hardware [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1320.917067] env[62692]: DEBUG nova.virt.hardware [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1320.917290] env[62692]: DEBUG nova.virt.hardware [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1320.917389] env[62692]: DEBUG nova.virt.hardware [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1320.917544] env[62692]: DEBUG nova.virt.hardware [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1320.917708] env[62692]: DEBUG nova.virt.hardware [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1320.917916] env[62692]: DEBUG nova.virt.hardware [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1320.918096] env[62692]: DEBUG nova.virt.hardware [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1320.918274] env[62692]: DEBUG nova.virt.hardware [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1320.918443] env[62692]: DEBUG nova.virt.hardware [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1320.918622] env[62692]: DEBUG nova.virt.hardware [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1320.919422] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bee3dd28-48c8-4d73-b9dd-67ee3fea1b67 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.925448] env[62692]: DEBUG oslo_vmware.api [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1320.925448] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]5283efc3-e1c8-a947-41c5-68fb45a5ca8e" [ 1320.925448] env[62692]: _type = "Task" [ 1320.925448] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1320.933303] env[62692]: DEBUG oslo_vmware.api [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5283efc3-e1c8-a947-41c5-68fb45a5ca8e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1321.436045] env[62692]: DEBUG oslo_vmware.api [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]5283efc3-e1c8-a947-41c5-68fb45a5ca8e, 'name': SearchDatastore_Task, 'duration_secs': 0.008962} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1321.441010] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Reconfiguring VM instance instance-00000078 to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1321.441280] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82c16dd2-7e13-4af1-98c5-985e07bd22a9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.458670] env[62692]: DEBUG oslo_vmware.api [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1321.458670] env[62692]: value = "task-1142202" [ 1321.458670] env[62692]: _type = "Task" [ 1321.458670] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1321.468597] env[62692]: DEBUG oslo_vmware.api [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142202, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1321.968949] env[62692]: DEBUG oslo_vmware.api [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142202, 'name': ReconfigVM_Task, 'duration_secs': 0.195904} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1321.971449] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Reconfigured VM instance instance-00000078 to detach disk 2000 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1321.971449] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca1e95a8-a753-4c9f-81ad-94fd848a711d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.994399] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Reconfiguring VM instance instance-00000078 to attach disk [datastore1] cb18c2e7-3241-4c9a-b1eb-a147b40530a1/cb18c2e7-3241-4c9a-b1eb-a147b40530a1.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1321.994655] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fff768c5-9b4f-407d-913e-54a721af5107 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.012416] env[62692]: DEBUG oslo_vmware.api [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1322.012416] env[62692]: value = "task-1142203" [ 1322.012416] env[62692]: _type = "Task" [ 1322.012416] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1322.023536] env[62692]: DEBUG oslo_vmware.api [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142203, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1322.522473] env[62692]: DEBUG oslo_vmware.api [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142203, 'name': ReconfigVM_Task, 'duration_secs': 0.304264} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1322.522770] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Reconfigured VM instance instance-00000078 to attach disk [datastore1] cb18c2e7-3241-4c9a-b1eb-a147b40530a1/cb18c2e7-3241-4c9a-b1eb-a147b40530a1.vmdk or device None with type thin {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1322.523603] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-273fbac3-beff-43a3-8cdb-41739449f460 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.543708] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36bb8dcf-69c3-4177-b4aa-5c9ace15fb7c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.565387] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-981a3997-e160-464e-bc26-158cf50e282d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.585535] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdfbe3ba-bd62-431f-bb0e-d563c815aae8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.592560] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1322.592812] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f3428916-67df-4f70-81c0-28416e544710 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.599258] env[62692]: DEBUG oslo_vmware.api [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1322.599258] env[62692]: value = "task-1142204" [ 1322.599258] env[62692]: _type = "Task" [ 1322.599258] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1322.606423] env[62692]: DEBUG oslo_vmware.api [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142204, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1323.109663] env[62692]: DEBUG oslo_vmware.api [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142204, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1323.150210] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "e4f1850a-5310-4ee7-a3b7-adbba64f5a83" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1323.150473] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "e4f1850a-5310-4ee7-a3b7-adbba64f5a83" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1323.150661] env[62692]: INFO nova.compute.manager [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Shelving [ 1323.611754] env[62692]: DEBUG oslo_vmware.api [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142204, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1323.657967] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1323.658197] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b7b01b70-deb7-4f13-bd13-79e5aec52af4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.664796] env[62692]: DEBUG oslo_vmware.api [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1323.664796] env[62692]: value = "task-1142205" [ 1323.664796] env[62692]: _type = "Task" [ 1323.664796] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1323.674122] env[62692]: DEBUG oslo_vmware.api [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142205, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1324.109927] env[62692]: DEBUG oslo_vmware.api [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142204, 'name': PowerOnVM_Task, 'duration_secs': 1.023164} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1324.110289] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1324.173801] env[62692]: DEBUG oslo_vmware.api [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142205, 'name': PowerOffVM_Task, 'duration_secs': 0.175456} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1324.174137] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1324.174953] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6691af8f-9212-4207-aa3e-b3223d2f9bfd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.193915] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-976819da-3168-49be-8e65-35788290d92e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.703193] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Creating Snapshot of the VM instance {{(pid=62692) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1324.704239] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-3f24aa72-de60-46a1-a738-259d6b99632c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.712558] env[62692]: DEBUG oslo_vmware.api [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1324.712558] env[62692]: value = "task-1142206" [ 1324.712558] env[62692]: _type = "Task" [ 1324.712558] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1324.721382] env[62692]: DEBUG oslo_vmware.api [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142206, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1325.155432] env[62692]: INFO nova.compute.manager [None req-e8d90504-f803-4098-99c5-008db8a814c4 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Updating instance to original state: 'active' [ 1325.222522] env[62692]: DEBUG oslo_vmware.api [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142206, 'name': CreateSnapshot_Task, 'duration_secs': 0.416431} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1325.222753] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Created Snapshot of the VM instance {{(pid=62692) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1325.223484] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1796524a-2222-40e7-a762-c3270d2c5411 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.740774] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Creating linked-clone VM from snapshot {{(pid=62692) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1325.741124] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-58466672-250b-41a7-b405-1982d21c6b08 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.751168] env[62692]: DEBUG oslo_vmware.api [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1325.751168] env[62692]: value = "task-1142207" [ 1325.751168] env[62692]: _type = "Task" [ 1325.751168] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1325.759560] env[62692]: DEBUG oslo_vmware.api [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142207, 'name': CloneVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1326.262440] env[62692]: DEBUG oslo_vmware.api [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142207, 'name': CloneVM_Task} progress is 94%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1326.444293] env[62692]: DEBUG oslo_concurrency.lockutils [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "cb18c2e7-3241-4c9a-b1eb-a147b40530a1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1326.444581] env[62692]: DEBUG oslo_concurrency.lockutils [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "cb18c2e7-3241-4c9a-b1eb-a147b40530a1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1326.444803] env[62692]: DEBUG oslo_concurrency.lockutils [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "cb18c2e7-3241-4c9a-b1eb-a147b40530a1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1326.444991] env[62692]: DEBUG oslo_concurrency.lockutils [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "cb18c2e7-3241-4c9a-b1eb-a147b40530a1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1326.445957] env[62692]: DEBUG oslo_concurrency.lockutils [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "cb18c2e7-3241-4c9a-b1eb-a147b40530a1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1326.448346] env[62692]: INFO nova.compute.manager [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Terminating instance [ 1326.450420] env[62692]: DEBUG nova.compute.manager [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1326.450764] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1326.450997] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-61dc803e-2bb5-4443-bcc9-0c631039a46a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.458185] env[62692]: DEBUG oslo_vmware.api [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1326.458185] env[62692]: value = "task-1142208" [ 1326.458185] env[62692]: _type = "Task" [ 1326.458185] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1326.466707] env[62692]: DEBUG oslo_vmware.api [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142208, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1326.761058] env[62692]: DEBUG oslo_vmware.api [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142207, 'name': CloneVM_Task} progress is 95%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1326.968254] env[62692]: DEBUG oslo_vmware.api [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142208, 'name': PowerOffVM_Task, 'duration_secs': 0.204136} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1326.968554] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1326.968760] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Volume detach. Driver type: vmdk {{(pid=62692) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1326.968961] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249104', 'volume_id': '77bf14fd-1e80-4e41-a6a9-f70f1356b51f', 'name': 'volume-77bf14fd-1e80-4e41-a6a9-f70f1356b51f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'cb18c2e7-3241-4c9a-b1eb-a147b40530a1', 'attached_at': '2024-10-05T20:41:44.000000', 'detached_at': '', 'volume_id': '77bf14fd-1e80-4e41-a6a9-f70f1356b51f', 'serial': '77bf14fd-1e80-4e41-a6a9-f70f1356b51f'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1326.969717] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0ab1c15-ca29-4c62-adba-abc3d22c572b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.990481] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d24a6bb-43de-4a84-879e-42adaaa05446 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.996851] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd338477-3509-43c3-a689-b8ecc4aa657a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.016495] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ea28c02-56ac-475b-ad39-affbfd25bae2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.030146] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] The volume has not been displaced from its original location: [datastore2] volume-77bf14fd-1e80-4e41-a6a9-f70f1356b51f/volume-77bf14fd-1e80-4e41-a6a9-f70f1356b51f.vmdk. No consolidation needed. {{(pid=62692) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1327.035352] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Reconfiguring VM instance instance-00000078 to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1327.035603] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b2fa70f-1551-474e-8238-cf6f022becda {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.053212] env[62692]: DEBUG oslo_vmware.api [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1327.053212] env[62692]: value = "task-1142209" [ 1327.053212] env[62692]: _type = "Task" [ 1327.053212] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1327.062723] env[62692]: DEBUG oslo_vmware.api [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142209, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1327.262056] env[62692]: DEBUG oslo_vmware.api [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142207, 'name': CloneVM_Task, 'duration_secs': 1.259866} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1327.262056] env[62692]: INFO nova.virt.vmwareapi.vmops [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Created linked-clone VM from snapshot [ 1327.262673] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f06c080d-208e-4097-bcb0-b84fa637b3a4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.269929] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Uploading image 081a5de5-caa9-4e2c-8a86-94deed626d62 {{(pid=62692) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1327.293177] env[62692]: DEBUG oslo_vmware.rw_handles [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1327.293177] env[62692]: value = "vm-249106" [ 1327.293177] env[62692]: _type = "VirtualMachine" [ 1327.293177] env[62692]: }. {{(pid=62692) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1327.293448] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-dd8dd291-cc82-41bf-bb24-18fe88268ebc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.300024] env[62692]: DEBUG oslo_vmware.rw_handles [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lease: (returnval){ [ 1327.300024] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52b486e9-74dc-6f37-e750-3b8eb3ffad4a" [ 1327.300024] env[62692]: _type = "HttpNfcLease" [ 1327.300024] env[62692]: } obtained for exporting VM: (result){ [ 1327.300024] env[62692]: value = "vm-249106" [ 1327.300024] env[62692]: _type = "VirtualMachine" [ 1327.300024] env[62692]: }. {{(pid=62692) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1327.300326] env[62692]: DEBUG oslo_vmware.api [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the lease: (returnval){ [ 1327.300326] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52b486e9-74dc-6f37-e750-3b8eb3ffad4a" [ 1327.300326] env[62692]: _type = "HttpNfcLease" [ 1327.300326] env[62692]: } to be ready. {{(pid=62692) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1327.306054] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1327.306054] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52b486e9-74dc-6f37-e750-3b8eb3ffad4a" [ 1327.306054] env[62692]: _type = "HttpNfcLease" [ 1327.306054] env[62692]: } is initializing. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1327.564698] env[62692]: DEBUG oslo_vmware.api [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142209, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1327.808888] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1327.808888] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52b486e9-74dc-6f37-e750-3b8eb3ffad4a" [ 1327.808888] env[62692]: _type = "HttpNfcLease" [ 1327.808888] env[62692]: } is ready. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1327.809417] env[62692]: DEBUG oslo_vmware.rw_handles [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1327.809417] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52b486e9-74dc-6f37-e750-3b8eb3ffad4a" [ 1327.809417] env[62692]: _type = "HttpNfcLease" [ 1327.809417] env[62692]: }. {{(pid=62692) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1327.810227] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67c6c28d-86ce-454a-b87a-10e57b38f6be {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.817339] env[62692]: DEBUG oslo_vmware.rw_handles [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520cce2d-6b69-a501-ebf7-0aa294736334/disk-0.vmdk from lease info. {{(pid=62692) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1327.817564] env[62692]: DEBUG oslo_vmware.rw_handles [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520cce2d-6b69-a501-ebf7-0aa294736334/disk-0.vmdk for reading. {{(pid=62692) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1327.905056] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6208699d-27f5-40d1-b29f-7b017a6c6460 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.063042] env[62692]: DEBUG oslo_vmware.api [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142209, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1328.564976] env[62692]: DEBUG oslo_vmware.api [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142209, 'name': ReconfigVM_Task, 'duration_secs': 1.273219} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1328.565541] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Reconfigured VM instance instance-00000078 to detach disk 2001 {{(pid=62692) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1328.570328] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3917782f-62b8-42c7-8774-cdd5b6ef90df {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.585722] env[62692]: DEBUG oslo_vmware.api [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1328.585722] env[62692]: value = "task-1142211" [ 1328.585722] env[62692]: _type = "Task" [ 1328.585722] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1328.593730] env[62692]: DEBUG oslo_vmware.api [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142211, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1329.096748] env[62692]: DEBUG oslo_vmware.api [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142211, 'name': ReconfigVM_Task, 'duration_secs': 0.145328} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1329.096748] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-249104', 'volume_id': '77bf14fd-1e80-4e41-a6a9-f70f1356b51f', 'name': 'volume-77bf14fd-1e80-4e41-a6a9-f70f1356b51f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'cb18c2e7-3241-4c9a-b1eb-a147b40530a1', 'attached_at': '2024-10-05T20:41:44.000000', 'detached_at': '', 'volume_id': '77bf14fd-1e80-4e41-a6a9-f70f1356b51f', 'serial': '77bf14fd-1e80-4e41-a6a9-f70f1356b51f'} {{(pid=62692) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1329.097074] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1329.097727] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c91e3d7a-7304-4424-a508-7eca7addc8b3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.104359] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1329.104626] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c0a73cb-4283-4f8c-81fb-1ff88b72d5ef {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.196461] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1329.196787] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1329.197065] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Deleting the datastore file [datastore1] cb18c2e7-3241-4c9a-b1eb-a147b40530a1 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1329.197387] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-24d58555-436f-41f8-8d3a-9fdc65b56de9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.204366] env[62692]: DEBUG oslo_vmware.api [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1329.204366] env[62692]: value = "task-1142213" [ 1329.204366] env[62692]: _type = "Task" [ 1329.204366] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1329.212887] env[62692]: DEBUG oslo_vmware.api [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142213, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1329.715116] env[62692]: DEBUG oslo_vmware.api [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142213, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.209608} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1329.715892] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1329.715892] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1329.715892] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1329.716672] env[62692]: INFO nova.compute.manager [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Took 3.27 seconds to destroy the instance on the hypervisor. [ 1329.716865] env[62692]: DEBUG oslo.service.loopingcall [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1329.717058] env[62692]: DEBUG nova.compute.manager [-] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1329.717160] env[62692]: DEBUG nova.network.neutron [-] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1330.182220] env[62692]: DEBUG nova.compute.manager [req-72650f04-5987-4668-ab59-b21f20797ce4 req-3a1f7e53-d59b-4475-9ce5-e6d37d7f7e66 service nova] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Received event network-vif-deleted-f84cd756-9d8a-4da6-bcd3-c377d3a0a875 {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1330.182448] env[62692]: INFO nova.compute.manager [req-72650f04-5987-4668-ab59-b21f20797ce4 req-3a1f7e53-d59b-4475-9ce5-e6d37d7f7e66 service nova] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Neutron deleted interface f84cd756-9d8a-4da6-bcd3-c377d3a0a875; detaching it from the instance and deleting it from the info cache [ 1330.182607] env[62692]: DEBUG nova.network.neutron [req-72650f04-5987-4668-ab59-b21f20797ce4 req-3a1f7e53-d59b-4475-9ce5-e6d37d7f7e66 service nova] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1330.660678] env[62692]: DEBUG nova.network.neutron [-] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1330.685782] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1a99bb6e-0b5a-41bc-8125-6ef5cf64520d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.695882] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54afcbe4-3807-4feb-9f7a-2cbdb2ee83bd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.723783] env[62692]: DEBUG nova.compute.manager [req-72650f04-5987-4668-ab59-b21f20797ce4 req-3a1f7e53-d59b-4475-9ce5-e6d37d7f7e66 service nova] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Detach interface failed, port_id=f84cd756-9d8a-4da6-bcd3-c377d3a0a875, reason: Instance cb18c2e7-3241-4c9a-b1eb-a147b40530a1 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1331.164220] env[62692]: INFO nova.compute.manager [-] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Took 1.45 seconds to deallocate network for instance. [ 1331.710120] env[62692]: INFO nova.compute.manager [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: cb18c2e7-3241-4c9a-b1eb-a147b40530a1] Took 0.55 seconds to detach 1 volumes for instance. [ 1332.216727] env[62692]: DEBUG oslo_concurrency.lockutils [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1332.217122] env[62692]: DEBUG oslo_concurrency.lockutils [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1332.217239] env[62692]: DEBUG oslo_concurrency.lockutils [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1332.240718] env[62692]: INFO nova.scheduler.client.report [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Deleted allocations for instance cb18c2e7-3241-4c9a-b1eb-a147b40530a1 [ 1332.748493] env[62692]: DEBUG oslo_concurrency.lockutils [None req-582e8b58-83a5-4968-966c-bd46d4fdfb31 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "cb18c2e7-3241-4c9a-b1eb-a147b40530a1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.304s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1333.819324] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "6e5b2bef-8bdf-402e-b4de-87a54c6d221b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1333.819662] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "6e5b2bef-8bdf-402e-b4de-87a54c6d221b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1334.322657] env[62692]: DEBUG nova.compute.manager [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Starting instance... {{(pid=62692) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1334.845285] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1334.845649] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1334.847168] env[62692]: INFO nova.compute.claims [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1335.143704] env[62692]: DEBUG oslo_vmware.rw_handles [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520cce2d-6b69-a501-ebf7-0aa294736334/disk-0.vmdk. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1335.144680] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c0cc250-d591-493c-89ac-fc2efc24d3c3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.151296] env[62692]: DEBUG oslo_vmware.rw_handles [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520cce2d-6b69-a501-ebf7-0aa294736334/disk-0.vmdk is in state: ready. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1335.151476] env[62692]: ERROR oslo_vmware.rw_handles [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520cce2d-6b69-a501-ebf7-0aa294736334/disk-0.vmdk due to incomplete transfer. [ 1335.151673] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-76d7e107-20c2-4564-a842-f1a6c4e6fb0b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.159679] env[62692]: DEBUG oslo_vmware.rw_handles [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520cce2d-6b69-a501-ebf7-0aa294736334/disk-0.vmdk. {{(pid=62692) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1335.159874] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Uploaded image 081a5de5-caa9-4e2c-8a86-94deed626d62 to the Glance image server {{(pid=62692) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1335.162234] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Destroying the VM {{(pid=62692) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1335.162489] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-2e31ddbf-cf22-40ac-886c-e4ba5e821264 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.168034] env[62692]: DEBUG oslo_vmware.api [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1335.168034] env[62692]: value = "task-1142214" [ 1335.168034] env[62692]: _type = "Task" [ 1335.168034] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1335.175431] env[62692]: DEBUG oslo_vmware.api [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142214, 'name': Destroy_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1335.678122] env[62692]: DEBUG oslo_vmware.api [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142214, 'name': Destroy_Task, 'duration_secs': 0.30572} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1335.678410] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Destroyed the VM [ 1335.678524] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Deleting Snapshot of the VM instance {{(pid=62692) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1335.678805] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-65feb31b-bcda-4b35-a4c6-0d562b8b60a6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.685024] env[62692]: DEBUG oslo_vmware.api [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1335.685024] env[62692]: value = "task-1142215" [ 1335.685024] env[62692]: _type = "Task" [ 1335.685024] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1335.692370] env[62692]: DEBUG oslo_vmware.api [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142215, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1335.872897] env[62692]: DEBUG nova.scheduler.client.report [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Refreshing inventories for resource provider 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1335.887101] env[62692]: DEBUG nova.scheduler.client.report [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Updating ProviderTree inventory for provider 470758a8-5efa-4875-aad5-f512a727752f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1335.887356] env[62692]: DEBUG nova.compute.provider_tree [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1335.897892] env[62692]: DEBUG nova.scheduler.client.report [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Refreshing aggregate associations for resource provider 470758a8-5efa-4875-aad5-f512a727752f, aggregates: None {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1335.914749] env[62692]: DEBUG nova.scheduler.client.report [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Refreshing trait associations for resource provider 470758a8-5efa-4875-aad5-f512a727752f, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1335.947162] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c9240c4-291c-49e9-a23d-aae481d94de2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.955657] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d220da-26b9-4cb3-a422-5afaab44b3f0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.984400] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ddca6e2-1689-40f9-b43b-3ba37f780c41 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.991136] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-871ca33f-d107-4ac5-982d-f25da53f09ec {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.003738] env[62692]: DEBUG nova.compute.provider_tree [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1336.194857] env[62692]: DEBUG oslo_vmware.api [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142215, 'name': RemoveSnapshot_Task, 'duration_secs': 0.395825} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1336.195195] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Deleted Snapshot of the VM instance {{(pid=62692) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1336.195483] env[62692]: DEBUG nova.compute.manager [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1336.196288] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69abe60d-f030-495c-b1fc-0684a6529380 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.506481] env[62692]: DEBUG nova.scheduler.client.report [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1336.708368] env[62692]: INFO nova.compute.manager [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Shelve offloading [ 1336.710140] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1336.710385] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cbd44142-6825-4955-84c9-31bb53269cb0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.719367] env[62692]: DEBUG oslo_vmware.api [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1336.719367] env[62692]: value = "task-1142216" [ 1336.719367] env[62692]: _type = "Task" [ 1336.719367] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1336.727229] env[62692]: DEBUG oslo_vmware.api [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142216, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1337.011026] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.165s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1337.011632] env[62692]: DEBUG nova.compute.manager [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Start building networks asynchronously for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1337.230836] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] VM already powered off {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1337.231065] env[62692]: DEBUG nova.compute.manager [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1337.231778] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f70c825-1eed-4e5f-a025-1a648b031ffe {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.237424] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "refresh_cache-e4f1850a-5310-4ee7-a3b7-adbba64f5a83" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1337.237590] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquired lock "refresh_cache-e4f1850a-5310-4ee7-a3b7-adbba64f5a83" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1337.237764] env[62692]: DEBUG nova.network.neutron [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1337.517064] env[62692]: DEBUG nova.compute.utils [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Using /dev/sd instead of None {{(pid=62692) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1337.518849] env[62692]: DEBUG nova.compute.manager [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Allocating IP information in the background. {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1337.519033] env[62692]: DEBUG nova.network.neutron [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] allocate_for_instance() {{(pid=62692) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1337.557393] env[62692]: DEBUG nova.policy [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6102acb4c19642d8a8190d828f5b5dfc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '697b4281872f4547ad3d051e644b95ce', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62692) authorize /opt/stack/nova/nova/policy.py:203}} [ 1337.827609] env[62692]: DEBUG nova.network.neutron [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Successfully created port: ef2af1cd-8e90-45c8-9f42-4fd1197a43dc {{(pid=62692) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1337.962964] env[62692]: DEBUG nova.network.neutron [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Updating instance_info_cache with network_info: [{"id": "1863d5f5-d859-4e5b-9d7f-b6e0df952e3d", "address": "fa:16:3e:3d:46:a6", "network": {"id": "22795f7f-18d7-4b10-8051-bcff4e0af3ac", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-445901728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96b921004f8a4bf68640c3810fe7b819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1863d5f5-d8", "ovs_interfaceid": "1863d5f5-d859-4e5b-9d7f-b6e0df952e3d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1338.022360] env[62692]: DEBUG nova.compute.manager [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Start building block device mappings for instance. {{(pid=62692) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1338.466116] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Releasing lock "refresh_cache-e4f1850a-5310-4ee7-a3b7-adbba64f5a83" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1338.783696] env[62692]: DEBUG nova.compute.manager [req-adfcf459-799f-4e65-958a-f74ca4079f79 req-680fca3e-877d-4588-9c3d-2e3abd186535 service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Received event network-vif-unplugged-1863d5f5-d859-4e5b-9d7f-b6e0df952e3d {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1338.783944] env[62692]: DEBUG oslo_concurrency.lockutils [req-adfcf459-799f-4e65-958a-f74ca4079f79 req-680fca3e-877d-4588-9c3d-2e3abd186535 service nova] Acquiring lock "e4f1850a-5310-4ee7-a3b7-adbba64f5a83-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1338.784244] env[62692]: DEBUG oslo_concurrency.lockutils [req-adfcf459-799f-4e65-958a-f74ca4079f79 req-680fca3e-877d-4588-9c3d-2e3abd186535 service nova] Lock "e4f1850a-5310-4ee7-a3b7-adbba64f5a83-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1338.784378] env[62692]: DEBUG oslo_concurrency.lockutils [req-adfcf459-799f-4e65-958a-f74ca4079f79 req-680fca3e-877d-4588-9c3d-2e3abd186535 service nova] Lock "e4f1850a-5310-4ee7-a3b7-adbba64f5a83-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1338.784553] env[62692]: DEBUG nova.compute.manager [req-adfcf459-799f-4e65-958a-f74ca4079f79 req-680fca3e-877d-4588-9c3d-2e3abd186535 service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] No waiting events found dispatching network-vif-unplugged-1863d5f5-d859-4e5b-9d7f-b6e0df952e3d {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1338.784727] env[62692]: WARNING nova.compute.manager [req-adfcf459-799f-4e65-958a-f74ca4079f79 req-680fca3e-877d-4588-9c3d-2e3abd186535 service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Received unexpected event network-vif-unplugged-1863d5f5-d859-4e5b-9d7f-b6e0df952e3d for instance with vm_state shelved and task_state shelving_offloading. [ 1338.876788] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1338.877723] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02e7395c-3f52-4001-a926-b7a05cbeae2d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.885679] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1338.885915] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6e8dacc3-7399-4b0c-aeb1-9b773647310c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.956123] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1338.956335] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1338.956525] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Deleting the datastore file [datastore2] e4f1850a-5310-4ee7-a3b7-adbba64f5a83 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1338.956814] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dd0baf74-6e9a-4a74-8792-29c16cfa57a5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.963453] env[62692]: DEBUG oslo_vmware.api [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1338.963453] env[62692]: value = "task-1142218" [ 1338.963453] env[62692]: _type = "Task" [ 1338.963453] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1338.971225] env[62692]: DEBUG oslo_vmware.api [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142218, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1339.031707] env[62692]: DEBUG nova.compute.manager [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Start spawning the instance on the hypervisor. {{(pid=62692) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1339.058591] env[62692]: DEBUG nova.virt.hardware [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-05T20:27:52Z,direct_url=,disk_format='vmdk',id=81eb64fb-5a0c-47b7-9948-7dcb01b911bd,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='989968e8796a4a70a4dbd045418a475f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-05T20:27:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1339.058883] env[62692]: DEBUG nova.virt.hardware [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1339.059078] env[62692]: DEBUG nova.virt.hardware [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1339.059312] env[62692]: DEBUG nova.virt.hardware [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1339.059493] env[62692]: DEBUG nova.virt.hardware [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1339.059665] env[62692]: DEBUG nova.virt.hardware [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1339.059905] env[62692]: DEBUG nova.virt.hardware [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1339.060106] env[62692]: DEBUG nova.virt.hardware [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1339.060302] env[62692]: DEBUG nova.virt.hardware [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1339.060501] env[62692]: DEBUG nova.virt.hardware [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1339.060709] env[62692]: DEBUG nova.virt.hardware [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1339.061627] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e343b77-01e2-4c71-b367-d1b8cffe8bbd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.069919] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad62512f-38ea-4b75-8798-96bdd44adde1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.196113] env[62692]: DEBUG nova.compute.manager [req-77ce3148-c32f-483a-896a-a8826207a57d req-f7c92f65-bdff-49d2-9cb4-ebb98c1f8503 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Received event network-vif-plugged-ef2af1cd-8e90-45c8-9f42-4fd1197a43dc {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1339.196398] env[62692]: DEBUG oslo_concurrency.lockutils [req-77ce3148-c32f-483a-896a-a8826207a57d req-f7c92f65-bdff-49d2-9cb4-ebb98c1f8503 service nova] Acquiring lock "6e5b2bef-8bdf-402e-b4de-87a54c6d221b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1339.196561] env[62692]: DEBUG oslo_concurrency.lockutils [req-77ce3148-c32f-483a-896a-a8826207a57d req-f7c92f65-bdff-49d2-9cb4-ebb98c1f8503 service nova] Lock "6e5b2bef-8bdf-402e-b4de-87a54c6d221b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1339.196730] env[62692]: DEBUG oslo_concurrency.lockutils [req-77ce3148-c32f-483a-896a-a8826207a57d req-f7c92f65-bdff-49d2-9cb4-ebb98c1f8503 service nova] Lock "6e5b2bef-8bdf-402e-b4de-87a54c6d221b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1339.196899] env[62692]: DEBUG nova.compute.manager [req-77ce3148-c32f-483a-896a-a8826207a57d req-f7c92f65-bdff-49d2-9cb4-ebb98c1f8503 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] No waiting events found dispatching network-vif-plugged-ef2af1cd-8e90-45c8-9f42-4fd1197a43dc {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1339.197080] env[62692]: WARNING nova.compute.manager [req-77ce3148-c32f-483a-896a-a8826207a57d req-f7c92f65-bdff-49d2-9cb4-ebb98c1f8503 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Received unexpected event network-vif-plugged-ef2af1cd-8e90-45c8-9f42-4fd1197a43dc for instance with vm_state building and task_state spawning. [ 1339.280837] env[62692]: DEBUG nova.network.neutron [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Successfully updated port: ef2af1cd-8e90-45c8-9f42-4fd1197a43dc {{(pid=62692) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1339.473646] env[62692]: DEBUG oslo_vmware.api [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142218, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16045} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1339.473852] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1339.474087] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1339.474295] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1339.498246] env[62692]: INFO nova.scheduler.client.report [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Deleted allocations for instance e4f1850a-5310-4ee7-a3b7-adbba64f5a83 [ 1339.783418] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "refresh_cache-6e5b2bef-8bdf-402e-b4de-87a54c6d221b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1339.783418] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquired lock "refresh_cache-6e5b2bef-8bdf-402e-b4de-87a54c6d221b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1339.783418] env[62692]: DEBUG nova.network.neutron [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1340.002892] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1340.003186] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1340.003423] env[62692]: DEBUG nova.objects.instance [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lazy-loading 'resources' on Instance uuid e4f1850a-5310-4ee7-a3b7-adbba64f5a83 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1340.324445] env[62692]: DEBUG nova.network.neutron [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Instance cache missing network info. {{(pid=62692) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1340.454181] env[62692]: DEBUG nova.network.neutron [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Updating instance_info_cache with network_info: [{"id": "ef2af1cd-8e90-45c8-9f42-4fd1197a43dc", "address": "fa:16:3e:5a:d0:c9", "network": {"id": "77695f27-3027-4a92-8b9b-0a90cdc53f62", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-337681783-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697b4281872f4547ad3d051e644b95ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef2af1cd-8e", "ovs_interfaceid": "ef2af1cd-8e90-45c8-9f42-4fd1197a43dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1340.506110] env[62692]: DEBUG nova.objects.instance [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lazy-loading 'numa_topology' on Instance uuid e4f1850a-5310-4ee7-a3b7-adbba64f5a83 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1340.808758] env[62692]: DEBUG nova.compute.manager [req-7420b64c-ce51-423e-b943-f640701d8b4a req-b3e6320a-a518-4f7c-93b4-55d6e66001fa service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Received event network-changed-1863d5f5-d859-4e5b-9d7f-b6e0df952e3d {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1340.808975] env[62692]: DEBUG nova.compute.manager [req-7420b64c-ce51-423e-b943-f640701d8b4a req-b3e6320a-a518-4f7c-93b4-55d6e66001fa service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Refreshing instance network info cache due to event network-changed-1863d5f5-d859-4e5b-9d7f-b6e0df952e3d. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1340.809219] env[62692]: DEBUG oslo_concurrency.lockutils [req-7420b64c-ce51-423e-b943-f640701d8b4a req-b3e6320a-a518-4f7c-93b4-55d6e66001fa service nova] Acquiring lock "refresh_cache-e4f1850a-5310-4ee7-a3b7-adbba64f5a83" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1340.809373] env[62692]: DEBUG oslo_concurrency.lockutils [req-7420b64c-ce51-423e-b943-f640701d8b4a req-b3e6320a-a518-4f7c-93b4-55d6e66001fa service nova] Acquired lock "refresh_cache-e4f1850a-5310-4ee7-a3b7-adbba64f5a83" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1340.809554] env[62692]: DEBUG nova.network.neutron [req-7420b64c-ce51-423e-b943-f640701d8b4a req-b3e6320a-a518-4f7c-93b4-55d6e66001fa service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Refreshing network info cache for port 1863d5f5-d859-4e5b-9d7f-b6e0df952e3d {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1340.957063] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Releasing lock "refresh_cache-6e5b2bef-8bdf-402e-b4de-87a54c6d221b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1340.957418] env[62692]: DEBUG nova.compute.manager [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Instance network_info: |[{"id": "ef2af1cd-8e90-45c8-9f42-4fd1197a43dc", "address": "fa:16:3e:5a:d0:c9", "network": {"id": "77695f27-3027-4a92-8b9b-0a90cdc53f62", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-337681783-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697b4281872f4547ad3d051e644b95ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef2af1cd-8e", "ovs_interfaceid": "ef2af1cd-8e90-45c8-9f42-4fd1197a43dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62692) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1340.957863] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5a:d0:c9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8e7f6f41-f4eb-4832-a390-730fca1cf717', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ef2af1cd-8e90-45c8-9f42-4fd1197a43dc', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1340.965515] env[62692]: DEBUG oslo.service.loopingcall [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1340.965726] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1340.965952] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-58fd5e47-014e-4e30-b19e-d9f6b6005167 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.987040] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1340.987040] env[62692]: value = "task-1142219" [ 1340.987040] env[62692]: _type = "Task" [ 1340.987040] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1340.995289] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1142219, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1341.008797] env[62692]: DEBUG nova.objects.base [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=62692) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1341.047741] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-878734e0-214d-4ddc-9695-acb550286685 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.055628] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5d695cd-f5d1-44fb-838d-51c8c7f0f1b1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.086862] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-def79f1b-e3b0-42d2-a058-3980ed40a1f8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.094258] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda5d99a-e1ff-4801-8c14-f123ea44c199 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.107630] env[62692]: DEBUG nova.compute.provider_tree [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1341.223653] env[62692]: DEBUG nova.compute.manager [req-a536bf37-d4e4-4465-bb25-a8823e63c66c req-28b6e821-55e1-4566-a72a-c2a8b57d65b2 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Received event network-changed-ef2af1cd-8e90-45c8-9f42-4fd1197a43dc {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1341.223899] env[62692]: DEBUG nova.compute.manager [req-a536bf37-d4e4-4465-bb25-a8823e63c66c req-28b6e821-55e1-4566-a72a-c2a8b57d65b2 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Refreshing instance network info cache due to event network-changed-ef2af1cd-8e90-45c8-9f42-4fd1197a43dc. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1341.224181] env[62692]: DEBUG oslo_concurrency.lockutils [req-a536bf37-d4e4-4465-bb25-a8823e63c66c req-28b6e821-55e1-4566-a72a-c2a8b57d65b2 service nova] Acquiring lock "refresh_cache-6e5b2bef-8bdf-402e-b4de-87a54c6d221b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1341.224372] env[62692]: DEBUG oslo_concurrency.lockutils [req-a536bf37-d4e4-4465-bb25-a8823e63c66c req-28b6e821-55e1-4566-a72a-c2a8b57d65b2 service nova] Acquired lock "refresh_cache-6e5b2bef-8bdf-402e-b4de-87a54c6d221b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1341.224559] env[62692]: DEBUG nova.network.neutron [req-a536bf37-d4e4-4465-bb25-a8823e63c66c req-28b6e821-55e1-4566-a72a-c2a8b57d65b2 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Refreshing network info cache for port ef2af1cd-8e90-45c8-9f42-4fd1197a43dc {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1341.498503] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1142219, 'name': CreateVM_Task, 'duration_secs': 0.27847} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1341.498840] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1341.499441] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1341.499731] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1341.500224] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1341.500624] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06f96397-f0bd-4a61-bb63-d2cafc5eb908 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.504880] env[62692]: DEBUG oslo_vmware.api [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1341.504880] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52709518-86df-4845-9ce3-96aa9dccc273" [ 1341.504880] env[62692]: _type = "Task" [ 1341.504880] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1341.511949] env[62692]: DEBUG oslo_vmware.api [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52709518-86df-4845-9ce3-96aa9dccc273, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1341.523700] env[62692]: DEBUG nova.network.neutron [req-7420b64c-ce51-423e-b943-f640701d8b4a req-b3e6320a-a518-4f7c-93b4-55d6e66001fa service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Updated VIF entry in instance network info cache for port 1863d5f5-d859-4e5b-9d7f-b6e0df952e3d. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1341.523988] env[62692]: DEBUG nova.network.neutron [req-7420b64c-ce51-423e-b943-f640701d8b4a req-b3e6320a-a518-4f7c-93b4-55d6e66001fa service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Updating instance_info_cache with network_info: [{"id": "1863d5f5-d859-4e5b-9d7f-b6e0df952e3d", "address": "fa:16:3e:3d:46:a6", "network": {"id": "22795f7f-18d7-4b10-8051-bcff4e0af3ac", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-445901728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96b921004f8a4bf68640c3810fe7b819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap1863d5f5-d8", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1341.610304] env[62692]: DEBUG nova.scheduler.client.report [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1341.796816] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "e4f1850a-5310-4ee7-a3b7-adbba64f5a83" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1341.905602] env[62692]: DEBUG nova.network.neutron [req-a536bf37-d4e4-4465-bb25-a8823e63c66c req-28b6e821-55e1-4566-a72a-c2a8b57d65b2 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Updated VIF entry in instance network info cache for port ef2af1cd-8e90-45c8-9f42-4fd1197a43dc. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1341.905964] env[62692]: DEBUG nova.network.neutron [req-a536bf37-d4e4-4465-bb25-a8823e63c66c req-28b6e821-55e1-4566-a72a-c2a8b57d65b2 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Updating instance_info_cache with network_info: [{"id": "ef2af1cd-8e90-45c8-9f42-4fd1197a43dc", "address": "fa:16:3e:5a:d0:c9", "network": {"id": "77695f27-3027-4a92-8b9b-0a90cdc53f62", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-337681783-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697b4281872f4547ad3d051e644b95ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef2af1cd-8e", "ovs_interfaceid": "ef2af1cd-8e90-45c8-9f42-4fd1197a43dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1342.016731] env[62692]: DEBUG oslo_vmware.api [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52709518-86df-4845-9ce3-96aa9dccc273, 'name': SearchDatastore_Task, 'duration_secs': 0.009537} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1342.017037] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1342.017277] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Processing image 81eb64fb-5a0c-47b7-9948-7dcb01b911bd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1342.017513] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1342.017686] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquired lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1342.017834] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1342.018116] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-29769e8f-4988-49f4-a2ac-accb2675a030 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.026818] env[62692]: DEBUG oslo_concurrency.lockutils [req-7420b64c-ce51-423e-b943-f640701d8b4a req-b3e6320a-a518-4f7c-93b4-55d6e66001fa service nova] Releasing lock "refresh_cache-e4f1850a-5310-4ee7-a3b7-adbba64f5a83" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1342.027209] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1342.027381] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1342.028076] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-516639ac-e555-4294-ac75-028f48919974 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.033060] env[62692]: DEBUG oslo_vmware.api [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1342.033060] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]528bb6ff-628d-8a7a-741d-af56740b726d" [ 1342.033060] env[62692]: _type = "Task" [ 1342.033060] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1342.040402] env[62692]: DEBUG oslo_vmware.api [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]528bb6ff-628d-8a7a-741d-af56740b726d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1342.114842] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.112s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1342.409550] env[62692]: DEBUG oslo_concurrency.lockutils [req-a536bf37-d4e4-4465-bb25-a8823e63c66c req-28b6e821-55e1-4566-a72a-c2a8b57d65b2 service nova] Releasing lock "refresh_cache-6e5b2bef-8bdf-402e-b4de-87a54c6d221b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1342.544109] env[62692]: DEBUG oslo_vmware.api [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]528bb6ff-628d-8a7a-741d-af56740b726d, 'name': SearchDatastore_Task, 'duration_secs': 0.008281} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1342.545049] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-389b9d59-e036-4caa-b9ae-3058a594febd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.550226] env[62692]: DEBUG oslo_vmware.api [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1342.550226] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52968375-2ff3-6351-ae24-c56afa3cad90" [ 1342.550226] env[62692]: _type = "Task" [ 1342.550226] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1342.557985] env[62692]: DEBUG oslo_vmware.api [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52968375-2ff3-6351-ae24-c56afa3cad90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1342.622602] env[62692]: DEBUG oslo_concurrency.lockutils [None req-29911c04-b50b-4d54-827c-33c5344617fa tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "e4f1850a-5310-4ee7-a3b7-adbba64f5a83" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 19.472s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1342.623367] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "e4f1850a-5310-4ee7-a3b7-adbba64f5a83" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.826s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1342.623367] env[62692]: INFO nova.compute.manager [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Unshelving [ 1343.061122] env[62692]: DEBUG oslo_vmware.api [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52968375-2ff3-6351-ae24-c56afa3cad90, 'name': SearchDatastore_Task, 'duration_secs': 0.009159} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1343.061391] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Releasing lock "[datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1343.061648] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 6e5b2bef-8bdf-402e-b4de-87a54c6d221b/6e5b2bef-8bdf-402e-b4de-87a54c6d221b.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1343.061898] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-88230ab6-0f69-4a32-aff1-b5d3fd4bcb4c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1343.069084] env[62692]: DEBUG oslo_vmware.api [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1343.069084] env[62692]: value = "task-1142220" [ 1343.069084] env[62692]: _type = "Task" [ 1343.069084] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1343.076232] env[62692]: DEBUG oslo_vmware.api [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142220, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1343.579325] env[62692]: DEBUG oslo_vmware.api [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142220, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.439924} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1343.579703] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/81eb64fb-5a0c-47b7-9948-7dcb01b911bd/81eb64fb-5a0c-47b7-9948-7dcb01b911bd.vmdk to [datastore2] 6e5b2bef-8bdf-402e-b4de-87a54c6d221b/6e5b2bef-8bdf-402e-b4de-87a54c6d221b.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1343.579810] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Extending root virtual disk to 1048576 {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1343.580070] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-574da7e5-2fd8-41c7-adb8-fd07b91212b1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1343.586066] env[62692]: DEBUG oslo_vmware.api [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1343.586066] env[62692]: value = "task-1142221" [ 1343.586066] env[62692]: _type = "Task" [ 1343.586066] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1343.593102] env[62692]: DEBUG oslo_vmware.api [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142221, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1343.646806] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1343.647088] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1343.647308] env[62692]: DEBUG nova.objects.instance [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lazy-loading 'pci_requests' on Instance uuid e4f1850a-5310-4ee7-a3b7-adbba64f5a83 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1344.096116] env[62692]: DEBUG oslo_vmware.api [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142221, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06072} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1344.096378] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Extended root virtual disk {{(pid=62692) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1344.097012] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c74d12f3-d2fa-41bf-8ec5-3fc858509a78 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.118301] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Reconfiguring VM instance instance-0000007a to attach disk [datastore2] 6e5b2bef-8bdf-402e-b4de-87a54c6d221b/6e5b2bef-8bdf-402e-b4de-87a54c6d221b.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1344.118540] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c7a2714a-41e8-43f0-829a-bcc44e29e51b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.137474] env[62692]: DEBUG oslo_vmware.api [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1344.137474] env[62692]: value = "task-1142222" [ 1344.137474] env[62692]: _type = "Task" [ 1344.137474] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1344.146055] env[62692]: DEBUG oslo_vmware.api [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142222, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1344.150736] env[62692]: DEBUG nova.objects.instance [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lazy-loading 'numa_topology' on Instance uuid e4f1850a-5310-4ee7-a3b7-adbba64f5a83 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1344.647727] env[62692]: DEBUG oslo_vmware.api [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142222, 'name': ReconfigVM_Task, 'duration_secs': 0.3212} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1344.649785] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Reconfigured VM instance instance-0000007a to attach disk [datastore2] 6e5b2bef-8bdf-402e-b4de-87a54c6d221b/6e5b2bef-8bdf-402e-b4de-87a54c6d221b.vmdk or device None with type sparse {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1344.649785] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0dbd48ff-f091-4fac-8ac6-b18c34f9e3e4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.653209] env[62692]: INFO nova.compute.claims [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1344.656570] env[62692]: DEBUG oslo_vmware.api [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1344.656570] env[62692]: value = "task-1142223" [ 1344.656570] env[62692]: _type = "Task" [ 1344.656570] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1344.664484] env[62692]: DEBUG oslo_vmware.api [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142223, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1345.168718] env[62692]: DEBUG oslo_vmware.api [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142223, 'name': Rename_Task, 'duration_secs': 0.144465} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1345.169501] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1345.169742] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ac8abebd-bf30-4733-a016-8effa12898f9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.176600] env[62692]: DEBUG oslo_vmware.api [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1345.176600] env[62692]: value = "task-1142224" [ 1345.176600] env[62692]: _type = "Task" [ 1345.176600] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1345.183749] env[62692]: DEBUG oslo_vmware.api [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142224, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1345.689387] env[62692]: DEBUG oslo_vmware.api [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142224, 'name': PowerOnVM_Task, 'duration_secs': 0.431543} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1345.689714] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1345.689867] env[62692]: INFO nova.compute.manager [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Took 6.66 seconds to spawn the instance on the hypervisor. [ 1345.690069] env[62692]: DEBUG nova.compute.manager [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1345.690863] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bcc55f2-eeaa-48c4-a01b-a6277566c3a8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.705661] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4602128-91ef-4530-9054-bf0336659165 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.712829] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c582dc60-f32d-489d-903c-3030c7e4b347 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.747750] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62edd23d-6870-4145-aa8d-7e6bb83aa0b7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.755773] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ed833fd-e374-45e2-918e-69f25f1d721f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.769205] env[62692]: DEBUG nova.compute.provider_tree [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1346.211479] env[62692]: INFO nova.compute.manager [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Took 11.38 seconds to build instance. [ 1346.272057] env[62692]: DEBUG nova.scheduler.client.report [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1346.714305] env[62692]: DEBUG oslo_concurrency.lockutils [None req-cc48ad14-9d0c-4a2b-ab4e-60dd5a2f35e2 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "6e5b2bef-8bdf-402e-b4de-87a54c6d221b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.894s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1346.777468] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.130s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1346.806985] env[62692]: INFO nova.network.neutron [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Updating port 1863d5f5-d859-4e5b-9d7f-b6e0df952e3d with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1347.171057] env[62692]: DEBUG nova.compute.manager [req-0445262d-2a1c-4021-a030-0acb6bc9c31f req-d4aacc47-1b81-4345-968e-76ab4336d226 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Received event network-changed-ef2af1cd-8e90-45c8-9f42-4fd1197a43dc {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1347.171057] env[62692]: DEBUG nova.compute.manager [req-0445262d-2a1c-4021-a030-0acb6bc9c31f req-d4aacc47-1b81-4345-968e-76ab4336d226 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Refreshing instance network info cache due to event network-changed-ef2af1cd-8e90-45c8-9f42-4fd1197a43dc. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1347.171337] env[62692]: DEBUG oslo_concurrency.lockutils [req-0445262d-2a1c-4021-a030-0acb6bc9c31f req-d4aacc47-1b81-4345-968e-76ab4336d226 service nova] Acquiring lock "refresh_cache-6e5b2bef-8bdf-402e-b4de-87a54c6d221b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1347.171428] env[62692]: DEBUG oslo_concurrency.lockutils [req-0445262d-2a1c-4021-a030-0acb6bc9c31f req-d4aacc47-1b81-4345-968e-76ab4336d226 service nova] Acquired lock "refresh_cache-6e5b2bef-8bdf-402e-b4de-87a54c6d221b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1347.171592] env[62692]: DEBUG nova.network.neutron [req-0445262d-2a1c-4021-a030-0acb6bc9c31f req-d4aacc47-1b81-4345-968e-76ab4336d226 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Refreshing network info cache for port ef2af1cd-8e90-45c8-9f42-4fd1197a43dc {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1347.885286] env[62692]: DEBUG nova.network.neutron [req-0445262d-2a1c-4021-a030-0acb6bc9c31f req-d4aacc47-1b81-4345-968e-76ab4336d226 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Updated VIF entry in instance network info cache for port ef2af1cd-8e90-45c8-9f42-4fd1197a43dc. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1347.885697] env[62692]: DEBUG nova.network.neutron [req-0445262d-2a1c-4021-a030-0acb6bc9c31f req-d4aacc47-1b81-4345-968e-76ab4336d226 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Updating instance_info_cache with network_info: [{"id": "ef2af1cd-8e90-45c8-9f42-4fd1197a43dc", "address": "fa:16:3e:5a:d0:c9", "network": {"id": "77695f27-3027-4a92-8b9b-0a90cdc53f62", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-337681783-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697b4281872f4547ad3d051e644b95ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef2af1cd-8e", "ovs_interfaceid": "ef2af1cd-8e90-45c8-9f42-4fd1197a43dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1348.208686] env[62692]: DEBUG nova.compute.manager [req-215eab79-94fd-44b8-940b-68a7e5c6b22d req-bdf7c403-59a3-4091-ae6c-6ac6485e5526 service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Received event network-vif-plugged-1863d5f5-d859-4e5b-9d7f-b6e0df952e3d {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1348.208952] env[62692]: DEBUG oslo_concurrency.lockutils [req-215eab79-94fd-44b8-940b-68a7e5c6b22d req-bdf7c403-59a3-4091-ae6c-6ac6485e5526 service nova] Acquiring lock "e4f1850a-5310-4ee7-a3b7-adbba64f5a83-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1348.209563] env[62692]: DEBUG oslo_concurrency.lockutils [req-215eab79-94fd-44b8-940b-68a7e5c6b22d req-bdf7c403-59a3-4091-ae6c-6ac6485e5526 service nova] Lock "e4f1850a-5310-4ee7-a3b7-adbba64f5a83-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1348.209796] env[62692]: DEBUG oslo_concurrency.lockutils [req-215eab79-94fd-44b8-940b-68a7e5c6b22d req-bdf7c403-59a3-4091-ae6c-6ac6485e5526 service nova] Lock "e4f1850a-5310-4ee7-a3b7-adbba64f5a83-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1348.210184] env[62692]: DEBUG nova.compute.manager [req-215eab79-94fd-44b8-940b-68a7e5c6b22d req-bdf7c403-59a3-4091-ae6c-6ac6485e5526 service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] No waiting events found dispatching network-vif-plugged-1863d5f5-d859-4e5b-9d7f-b6e0df952e3d {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1348.210460] env[62692]: WARNING nova.compute.manager [req-215eab79-94fd-44b8-940b-68a7e5c6b22d req-bdf7c403-59a3-4091-ae6c-6ac6485e5526 service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Received unexpected event network-vif-plugged-1863d5f5-d859-4e5b-9d7f-b6e0df952e3d for instance with vm_state shelved_offloaded and task_state spawning. [ 1348.298411] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "refresh_cache-e4f1850a-5310-4ee7-a3b7-adbba64f5a83" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1348.298605] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquired lock "refresh_cache-e4f1850a-5310-4ee7-a3b7-adbba64f5a83" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1348.298948] env[62692]: DEBUG nova.network.neutron [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1348.388878] env[62692]: DEBUG oslo_concurrency.lockutils [req-0445262d-2a1c-4021-a030-0acb6bc9c31f req-d4aacc47-1b81-4345-968e-76ab4336d226 service nova] Releasing lock "refresh_cache-6e5b2bef-8bdf-402e-b4de-87a54c6d221b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1348.998142] env[62692]: DEBUG nova.network.neutron [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Updating instance_info_cache with network_info: [{"id": "1863d5f5-d859-4e5b-9d7f-b6e0df952e3d", "address": "fa:16:3e:3d:46:a6", "network": {"id": "22795f7f-18d7-4b10-8051-bcff4e0af3ac", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-445901728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96b921004f8a4bf68640c3810fe7b819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1863d5f5-d8", "ovs_interfaceid": "1863d5f5-d859-4e5b-9d7f-b6e0df952e3d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1349.500590] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Releasing lock "refresh_cache-e4f1850a-5310-4ee7-a3b7-adbba64f5a83" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1349.527348] env[62692]: DEBUG nova.virt.hardware [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='7eff121a870e94165046d0431670dbb8',container_format='bare',created_at=2024-10-05T20:41:47Z,direct_url=,disk_format='vmdk',id=081a5de5-caa9-4e2c-8a86-94deed626d62,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-129342005-shelved',owner='96b921004f8a4bf68640c3810fe7b819',properties=ImageMetaProps,protected=,size=31668224,status='active',tags=,updated_at=2024-10-05T20:42:00Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1349.527621] env[62692]: DEBUG nova.virt.hardware [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1349.527789] env[62692]: DEBUG nova.virt.hardware [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1349.527980] env[62692]: DEBUG nova.virt.hardware [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1349.528154] env[62692]: DEBUG nova.virt.hardware [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1349.528532] env[62692]: DEBUG nova.virt.hardware [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1349.528532] env[62692]: DEBUG nova.virt.hardware [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1349.528681] env[62692]: DEBUG nova.virt.hardware [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1349.528858] env[62692]: DEBUG nova.virt.hardware [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1349.529038] env[62692]: DEBUG nova.virt.hardware [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1349.529225] env[62692]: DEBUG nova.virt.hardware [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1349.530101] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab88c05e-e5a9-43f3-9a3d-4975b3d087d8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.538617] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-942a64c4-e520-476d-8933-e7adfc0a71ed {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.552995] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3d:46:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd1da5fc2-0280-4f76-ac97-20ea4bc7bb16', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1863d5f5-d859-4e5b-9d7f-b6e0df952e3d', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1349.560411] env[62692]: DEBUG oslo.service.loopingcall [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1349.560629] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1349.560847] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8140f4d3-b902-4212-bb65-4c927eef507b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.579470] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1349.579470] env[62692]: value = "task-1142225" [ 1349.579470] env[62692]: _type = "Task" [ 1349.579470] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1349.586700] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1142225, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1350.089723] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1142225, 'name': CreateVM_Task, 'duration_secs': 0.310653} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1350.090029] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1350.090513] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/081a5de5-caa9-4e2c-8a86-94deed626d62" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1350.090680] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquired lock "[datastore1] devstack-image-cache_base/081a5de5-caa9-4e2c-8a86-94deed626d62" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1350.091066] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/081a5de5-caa9-4e2c-8a86-94deed626d62" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1350.091318] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2e3060f-1672-4082-a5dd-ec6c9d2ed51c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.095582] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1350.095582] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52bf4820-2439-5286-9989-3182a1a02dfe" [ 1350.095582] env[62692]: _type = "Task" [ 1350.095582] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1350.102774] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52bf4820-2439-5286-9989-3182a1a02dfe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1350.233961] env[62692]: DEBUG nova.compute.manager [req-7d84bb74-2e28-4574-bf2d-80403caa6c4a req-06f3dac6-8c2b-4da3-8edf-00e22252b31a service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Received event network-changed-1863d5f5-d859-4e5b-9d7f-b6e0df952e3d {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1350.234224] env[62692]: DEBUG nova.compute.manager [req-7d84bb74-2e28-4574-bf2d-80403caa6c4a req-06f3dac6-8c2b-4da3-8edf-00e22252b31a service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Refreshing instance network info cache due to event network-changed-1863d5f5-d859-4e5b-9d7f-b6e0df952e3d. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1350.234447] env[62692]: DEBUG oslo_concurrency.lockutils [req-7d84bb74-2e28-4574-bf2d-80403caa6c4a req-06f3dac6-8c2b-4da3-8edf-00e22252b31a service nova] Acquiring lock "refresh_cache-e4f1850a-5310-4ee7-a3b7-adbba64f5a83" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1350.234597] env[62692]: DEBUG oslo_concurrency.lockutils [req-7d84bb74-2e28-4574-bf2d-80403caa6c4a req-06f3dac6-8c2b-4da3-8edf-00e22252b31a service nova] Acquired lock "refresh_cache-e4f1850a-5310-4ee7-a3b7-adbba64f5a83" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1350.234766] env[62692]: DEBUG nova.network.neutron [req-7d84bb74-2e28-4574-bf2d-80403caa6c4a req-06f3dac6-8c2b-4da3-8edf-00e22252b31a service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Refreshing network info cache for port 1863d5f5-d859-4e5b-9d7f-b6e0df952e3d {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1350.606163] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Releasing lock "[datastore1] devstack-image-cache_base/081a5de5-caa9-4e2c-8a86-94deed626d62" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1350.606449] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Processing image 081a5de5-caa9-4e2c-8a86-94deed626d62 {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1350.606692] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/081a5de5-caa9-4e2c-8a86-94deed626d62/081a5de5-caa9-4e2c-8a86-94deed626d62.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1350.606850] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquired lock "[datastore1] devstack-image-cache_base/081a5de5-caa9-4e2c-8a86-94deed626d62/081a5de5-caa9-4e2c-8a86-94deed626d62.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1350.607048] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1350.607323] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-af55f6d3-dbb6-4d17-b387-bd425814afe7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.615619] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1350.615798] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1350.616544] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7daa24d-8209-4368-ad66-dedc76722b8d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.621755] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1350.621755] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52bb425a-6123-526a-effc-2d6efc2578d6" [ 1350.621755] env[62692]: _type = "Task" [ 1350.621755] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1350.631627] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52bb425a-6123-526a-effc-2d6efc2578d6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1350.938080] env[62692]: DEBUG nova.network.neutron [req-7d84bb74-2e28-4574-bf2d-80403caa6c4a req-06f3dac6-8c2b-4da3-8edf-00e22252b31a service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Updated VIF entry in instance network info cache for port 1863d5f5-d859-4e5b-9d7f-b6e0df952e3d. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1350.938486] env[62692]: DEBUG nova.network.neutron [req-7d84bb74-2e28-4574-bf2d-80403caa6c4a req-06f3dac6-8c2b-4da3-8edf-00e22252b31a service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Updating instance_info_cache with network_info: [{"id": "1863d5f5-d859-4e5b-9d7f-b6e0df952e3d", "address": "fa:16:3e:3d:46:a6", "network": {"id": "22795f7f-18d7-4b10-8051-bcff4e0af3ac", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-445901728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96b921004f8a4bf68640c3810fe7b819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1863d5f5-d8", "ovs_interfaceid": "1863d5f5-d859-4e5b-9d7f-b6e0df952e3d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1351.131521] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Preparing fetch location {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1351.131878] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Fetch image to [datastore1] OSTACK_IMG_0169f842-45c0-4e72-9b67-fe9e5d84b6ba/OSTACK_IMG_0169f842-45c0-4e72-9b67-fe9e5d84b6ba.vmdk {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1351.131991] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Downloading stream optimized image 081a5de5-caa9-4e2c-8a86-94deed626d62 to [datastore1] OSTACK_IMG_0169f842-45c0-4e72-9b67-fe9e5d84b6ba/OSTACK_IMG_0169f842-45c0-4e72-9b67-fe9e5d84b6ba.vmdk on the data store datastore1 as vApp {{(pid=62692) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1351.132223] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Downloading image file data 081a5de5-caa9-4e2c-8a86-94deed626d62 to the ESX as VM named 'OSTACK_IMG_0169f842-45c0-4e72-9b67-fe9e5d84b6ba' {{(pid=62692) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1351.195908] env[62692]: DEBUG oslo_vmware.rw_handles [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1351.195908] env[62692]: value = "resgroup-9" [ 1351.195908] env[62692]: _type = "ResourcePool" [ 1351.195908] env[62692]: }. {{(pid=62692) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1351.196203] env[62692]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-7d284311-5da7-4363-a35e-44e093a118bb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1351.216632] env[62692]: DEBUG oslo_vmware.rw_handles [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lease: (returnval){ [ 1351.216632] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52b11b7a-0bf2-94de-3e54-f13a6f7f58bc" [ 1351.216632] env[62692]: _type = "HttpNfcLease" [ 1351.216632] env[62692]: } obtained for vApp import into resource pool (val){ [ 1351.216632] env[62692]: value = "resgroup-9" [ 1351.216632] env[62692]: _type = "ResourcePool" [ 1351.216632] env[62692]: }. {{(pid=62692) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1351.217021] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the lease: (returnval){ [ 1351.217021] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52b11b7a-0bf2-94de-3e54-f13a6f7f58bc" [ 1351.217021] env[62692]: _type = "HttpNfcLease" [ 1351.217021] env[62692]: } to be ready. {{(pid=62692) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1351.222756] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1351.222756] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52b11b7a-0bf2-94de-3e54-f13a6f7f58bc" [ 1351.222756] env[62692]: _type = "HttpNfcLease" [ 1351.222756] env[62692]: } is initializing. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1351.441499] env[62692]: DEBUG oslo_concurrency.lockutils [req-7d84bb74-2e28-4574-bf2d-80403caa6c4a req-06f3dac6-8c2b-4da3-8edf-00e22252b31a service nova] Releasing lock "refresh_cache-e4f1850a-5310-4ee7-a3b7-adbba64f5a83" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1351.725478] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1351.725478] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52b11b7a-0bf2-94de-3e54-f13a6f7f58bc" [ 1351.725478] env[62692]: _type = "HttpNfcLease" [ 1351.725478] env[62692]: } is ready. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1351.725820] env[62692]: DEBUG oslo_vmware.rw_handles [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1351.725820] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52b11b7a-0bf2-94de-3e54-f13a6f7f58bc" [ 1351.725820] env[62692]: _type = "HttpNfcLease" [ 1351.725820] env[62692]: }. {{(pid=62692) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1351.726552] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9999093-3c26-4b16-ba98-5a3d7e72712d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1351.733235] env[62692]: DEBUG oslo_vmware.rw_handles [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525e201b-399b-f842-7a03-b481d8ce7ad3/disk-0.vmdk from lease info. {{(pid=62692) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1351.733417] env[62692]: DEBUG oslo_vmware.rw_handles [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Creating HTTP connection to write to file with size = 31668224 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525e201b-399b-f842-7a03-b481d8ce7ad3/disk-0.vmdk. {{(pid=62692) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1351.798223] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8cc13416-a848-4761-aa15-7ae023ad072d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.932891] env[62692]: DEBUG oslo_vmware.rw_handles [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Completed reading data from the image iterator. {{(pid=62692) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1352.933359] env[62692]: DEBUG oslo_vmware.rw_handles [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525e201b-399b-f842-7a03-b481d8ce7ad3/disk-0.vmdk. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1352.934042] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bf344f4-98d9-419b-b061-10f5889ff1b8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.940570] env[62692]: DEBUG oslo_vmware.rw_handles [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525e201b-399b-f842-7a03-b481d8ce7ad3/disk-0.vmdk is in state: ready. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1352.940851] env[62692]: DEBUG oslo_vmware.rw_handles [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525e201b-399b-f842-7a03-b481d8ce7ad3/disk-0.vmdk. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1352.940981] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-471b7fff-ac25-4e7f-a3ac-ece673bfd64e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.143810] env[62692]: DEBUG oslo_vmware.rw_handles [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525e201b-399b-f842-7a03-b481d8ce7ad3/disk-0.vmdk. {{(pid=62692) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1353.144050] env[62692]: INFO nova.virt.vmwareapi.images [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Downloaded image file data 081a5de5-caa9-4e2c-8a86-94deed626d62 [ 1353.145039] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-286c97af-edd0-4bc5-984c-cb174a6edf7c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.160242] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-85a9c0d6-7198-4719-9e8f-d0d711287214 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.189367] env[62692]: INFO nova.virt.vmwareapi.images [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] The imported VM was unregistered [ 1353.191879] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Caching image {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1353.192141] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Creating directory with path [datastore1] devstack-image-cache_base/081a5de5-caa9-4e2c-8a86-94deed626d62 {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1353.192406] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dfbb2bc7-7d18-46f8-8631-15e2149c86f8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.213377] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Created directory with path [datastore1] devstack-image-cache_base/081a5de5-caa9-4e2c-8a86-94deed626d62 {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1353.213570] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_0169f842-45c0-4e72-9b67-fe9e5d84b6ba/OSTACK_IMG_0169f842-45c0-4e72-9b67-fe9e5d84b6ba.vmdk to [datastore1] devstack-image-cache_base/081a5de5-caa9-4e2c-8a86-94deed626d62/081a5de5-caa9-4e2c-8a86-94deed626d62.vmdk. {{(pid=62692) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1353.213826] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-ddea7488-702d-4d21-9e38-c972e16f62c1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.220618] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1353.220618] env[62692]: value = "task-1142228" [ 1353.220618] env[62692]: _type = "Task" [ 1353.220618] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1353.228735] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142228, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1353.731141] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142228, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1354.231447] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142228, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1354.381506] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1354.381747] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1354.381905] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Starting heal instance info cache {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1354.382061] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Rebuilding the list of instances to heal {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1354.732311] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142228, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1354.885886] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "refresh_cache-e4f1850a-5310-4ee7-a3b7-adbba64f5a83" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1354.886135] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquired lock "refresh_cache-e4f1850a-5310-4ee7-a3b7-adbba64f5a83" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1354.886199] env[62692]: DEBUG nova.network.neutron [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Forcefully refreshing network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1354.886350] env[62692]: DEBUG nova.objects.instance [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lazy-loading 'info_cache' on Instance uuid e4f1850a-5310-4ee7-a3b7-adbba64f5a83 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1355.232542] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142228, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1355.732897] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142228, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.206295} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1355.733723] env[62692]: INFO nova.virt.vmwareapi.ds_util [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_0169f842-45c0-4e72-9b67-fe9e5d84b6ba/OSTACK_IMG_0169f842-45c0-4e72-9b67-fe9e5d84b6ba.vmdk to [datastore1] devstack-image-cache_base/081a5de5-caa9-4e2c-8a86-94deed626d62/081a5de5-caa9-4e2c-8a86-94deed626d62.vmdk. [ 1355.733723] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Cleaning up location [datastore1] OSTACK_IMG_0169f842-45c0-4e72-9b67-fe9e5d84b6ba {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1355.733878] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_0169f842-45c0-4e72-9b67-fe9e5d84b6ba {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1355.734019] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-861ce573-dec8-42f0-b674-c9f44050edd7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.740222] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1355.740222] env[62692]: value = "task-1142229" [ 1355.740222] env[62692]: _type = "Task" [ 1355.740222] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1355.747183] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142229, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1356.249840] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142229, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.035827} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1356.250171] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1356.250300] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Releasing lock "[datastore1] devstack-image-cache_base/081a5de5-caa9-4e2c-8a86-94deed626d62/081a5de5-caa9-4e2c-8a86-94deed626d62.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1356.250546] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/081a5de5-caa9-4e2c-8a86-94deed626d62/081a5de5-caa9-4e2c-8a86-94deed626d62.vmdk to [datastore1] e4f1850a-5310-4ee7-a3b7-adbba64f5a83/e4f1850a-5310-4ee7-a3b7-adbba64f5a83.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1356.250795] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4bfd7e28-f130-4476-96d5-c64446273251 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.258418] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1356.258418] env[62692]: value = "task-1142230" [ 1356.258418] env[62692]: _type = "Task" [ 1356.258418] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1356.265722] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142230, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1356.607192] env[62692]: DEBUG nova.network.neutron [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Updating instance_info_cache with network_info: [{"id": "1863d5f5-d859-4e5b-9d7f-b6e0df952e3d", "address": "fa:16:3e:3d:46:a6", "network": {"id": "22795f7f-18d7-4b10-8051-bcff4e0af3ac", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-445901728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96b921004f8a4bf68640c3810fe7b819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1863d5f5-d8", "ovs_interfaceid": "1863d5f5-d859-4e5b-9d7f-b6e0df952e3d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1356.768282] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142230, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1357.110088] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Releasing lock "refresh_cache-e4f1850a-5310-4ee7-a3b7-adbba64f5a83" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1357.110352] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Updated the network info_cache for instance {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1357.110520] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1357.110697] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1357.110853] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1357.111025] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1357.111161] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1357.111310] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1357.111441] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62692) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1357.111592] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager.update_available_resource {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1357.270061] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142230, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1357.614913] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1357.615168] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1357.615369] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1357.615533] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62692) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1357.616570] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a03c5001-3401-42c7-a9d2-abd3326a3396 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.625045] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94fbf198-dca9-468a-a21f-7a0ecc0f9fef {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.639835] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-403ea7e1-b8ea-419f-a1af-5a78f08328bd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.646644] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e702839c-3d8f-4d8e-bcd0-53478665f4df {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.675774] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181055MB free_disk=166GB free_vcpus=48 pci_devices=None {{(pid=62692) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1357.676091] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1357.676180] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1357.771645] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142230, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1358.271780] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142230, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1358.701147] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 6e5b2bef-8bdf-402e-b4de-87a54c6d221b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1358.701326] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance e4f1850a-5310-4ee7-a3b7-adbba64f5a83 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1358.701508] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=62692) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1358.701658] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=62692) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1358.735018] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ed26d37-3a4e-4f30-9b37-445fc43b86d8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.742398] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4588473b-b7a9-4953-8cd6-fa2f6ef879da {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.774089] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c55c7e1b-68b4-4804-a50f-fac4ded14868 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.783178] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142230, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.289553} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1358.784341] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00fddd38-0a42-4fd7-90e4-b1d336771862 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.787753] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/081a5de5-caa9-4e2c-8a86-94deed626d62/081a5de5-caa9-4e2c-8a86-94deed626d62.vmdk to [datastore1] e4f1850a-5310-4ee7-a3b7-adbba64f5a83/e4f1850a-5310-4ee7-a3b7-adbba64f5a83.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1358.788460] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa2caeac-288a-4d04-b914-d51e4e435901 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.800030] env[62692]: DEBUG nova.compute.provider_tree [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1358.817988] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Reconfiguring VM instance instance-00000079 to attach disk [datastore1] e4f1850a-5310-4ee7-a3b7-adbba64f5a83/e4f1850a-5310-4ee7-a3b7-adbba64f5a83.vmdk or device None with type streamOptimized {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1358.818905] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0786d2f-21ce-406d-a1a6-b77f034d8296 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.838766] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1358.838766] env[62692]: value = "task-1142231" [ 1358.838766] env[62692]: _type = "Task" [ 1358.838766] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1358.847862] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142231, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1359.319507] env[62692]: DEBUG nova.scheduler.client.report [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1359.348417] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142231, 'name': ReconfigVM_Task, 'duration_secs': 0.485239} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1359.348690] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Reconfigured VM instance instance-00000079 to attach disk [datastore1] e4f1850a-5310-4ee7-a3b7-adbba64f5a83/e4f1850a-5310-4ee7-a3b7-adbba64f5a83.vmdk or device None with type streamOptimized {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1359.349323] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2fee12cf-9620-4d71-ac93-3320df52f0f2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.355419] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1359.355419] env[62692]: value = "task-1142232" [ 1359.355419] env[62692]: _type = "Task" [ 1359.355419] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1359.362683] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142232, 'name': Rename_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1359.824717] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62692) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1359.824717] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.148s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1359.866108] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142232, 'name': Rename_Task, 'duration_secs': 0.330011} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1359.866108] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1359.866108] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-edc18b9e-b81a-40d4-bcec-43d1a5f54cdb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.871789] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1359.871789] env[62692]: value = "task-1142233" [ 1359.871789] env[62692]: _type = "Task" [ 1359.871789] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1359.878873] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142233, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1360.381116] env[62692]: DEBUG oslo_vmware.api [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142233, 'name': PowerOnVM_Task, 'duration_secs': 0.438908} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1360.381511] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1360.476174] env[62692]: DEBUG nova.compute.manager [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1360.477157] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7618eaf1-ac85-4407-a2bc-3d7816666088 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.994456] env[62692]: DEBUG oslo_concurrency.lockutils [None req-7e8bd2f8-60b1-4dff-aeed-2ba15b849e16 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "e4f1850a-5310-4ee7-a3b7-adbba64f5a83" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 18.371s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1384.916040] env[62692]: DEBUG oslo_concurrency.lockutils [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "6e5b2bef-8bdf-402e-b4de-87a54c6d221b" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1384.916451] env[62692]: DEBUG oslo_concurrency.lockutils [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "6e5b2bef-8bdf-402e-b4de-87a54c6d221b" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1384.916603] env[62692]: INFO nova.compute.manager [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Shelving [ 1385.423705] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1385.423967] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-89f94116-3b3b-4118-aceb-3fb08a8bd20f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.431460] env[62692]: DEBUG oslo_vmware.api [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1385.431460] env[62692]: value = "task-1142234" [ 1385.431460] env[62692]: _type = "Task" [ 1385.431460] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1385.439457] env[62692]: DEBUG oslo_vmware.api [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142234, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1385.941170] env[62692]: DEBUG oslo_vmware.api [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142234, 'name': PowerOffVM_Task, 'duration_secs': 0.171086} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1385.941644] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1385.942453] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd0b17e8-14e0-40fe-81ac-07090930040f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.960458] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52d685ae-c5d0-49ec-b7b9-16930af11bd6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1386.470231] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Creating Snapshot of the VM instance {{(pid=62692) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1386.470561] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-46e4c292-daaa-4160-8b0b-bbc94766fc64 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1386.479399] env[62692]: DEBUG oslo_vmware.api [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1386.479399] env[62692]: value = "task-1142235" [ 1386.479399] env[62692]: _type = "Task" [ 1386.479399] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1386.487854] env[62692]: DEBUG oslo_vmware.api [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142235, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1386.989846] env[62692]: DEBUG oslo_vmware.api [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142235, 'name': CreateSnapshot_Task, 'duration_secs': 0.429002} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1386.990335] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Created Snapshot of the VM instance {{(pid=62692) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1386.991076] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5736ef97-4577-42a5-bb4f-6d4546f08538 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1387.508404] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Creating linked-clone VM from snapshot {{(pid=62692) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1387.508725] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d65768d2-9e86-4cc5-bb39-0124cf5cc4d1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1387.518658] env[62692]: DEBUG oslo_vmware.api [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1387.518658] env[62692]: value = "task-1142236" [ 1387.518658] env[62692]: _type = "Task" [ 1387.518658] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1387.526128] env[62692]: DEBUG oslo_vmware.api [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142236, 'name': CloneVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1388.028711] env[62692]: DEBUG oslo_vmware.api [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142236, 'name': CloneVM_Task} progress is 94%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1388.528310] env[62692]: DEBUG oslo_vmware.api [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142236, 'name': CloneVM_Task} progress is 100%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1389.031048] env[62692]: DEBUG oslo_vmware.api [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142236, 'name': CloneVM_Task, 'duration_secs': 1.042717} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1389.031048] env[62692]: INFO nova.virt.vmwareapi.vmops [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Created linked-clone VM from snapshot [ 1389.031394] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b92a5be-c3e6-47a7-b8df-b2d2130e2ac4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.039712] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Uploading image 8900c02f-8bce-4dc2-b362-b565762af1fd {{(pid=62692) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1389.059251] env[62692]: DEBUG oslo_vmware.rw_handles [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1389.059251] env[62692]: value = "vm-249111" [ 1389.059251] env[62692]: _type = "VirtualMachine" [ 1389.059251] env[62692]: }. {{(pid=62692) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1389.059500] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-ed18f5b1-38ec-4907-8680-d5fd8b80f7ff {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.065882] env[62692]: DEBUG oslo_vmware.rw_handles [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lease: (returnval){ [ 1389.065882] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]523a8a52-050d-fd06-402c-8bcde303b561" [ 1389.065882] env[62692]: _type = "HttpNfcLease" [ 1389.065882] env[62692]: } obtained for exporting VM: (result){ [ 1389.065882] env[62692]: value = "vm-249111" [ 1389.065882] env[62692]: _type = "VirtualMachine" [ 1389.065882] env[62692]: }. {{(pid=62692) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1389.066243] env[62692]: DEBUG oslo_vmware.api [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the lease: (returnval){ [ 1389.066243] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]523a8a52-050d-fd06-402c-8bcde303b561" [ 1389.066243] env[62692]: _type = "HttpNfcLease" [ 1389.066243] env[62692]: } to be ready. {{(pid=62692) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1389.071661] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1389.071661] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]523a8a52-050d-fd06-402c-8bcde303b561" [ 1389.071661] env[62692]: _type = "HttpNfcLease" [ 1389.071661] env[62692]: } is initializing. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1389.575352] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1389.575352] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]523a8a52-050d-fd06-402c-8bcde303b561" [ 1389.575352] env[62692]: _type = "HttpNfcLease" [ 1389.575352] env[62692]: } is ready. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1389.575756] env[62692]: DEBUG oslo_vmware.rw_handles [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1389.575756] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]523a8a52-050d-fd06-402c-8bcde303b561" [ 1389.575756] env[62692]: _type = "HttpNfcLease" [ 1389.575756] env[62692]: }. {{(pid=62692) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1389.576405] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76d3cf7a-f49c-4ba7-b55f-e64c1a2938fc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.583539] env[62692]: DEBUG oslo_vmware.rw_handles [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526b55f8-d151-44c8-d576-e48b1c8fdeac/disk-0.vmdk from lease info. {{(pid=62692) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1389.583714] env[62692]: DEBUG oslo_vmware.rw_handles [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526b55f8-d151-44c8-d576-e48b1c8fdeac/disk-0.vmdk for reading. {{(pid=62692) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1389.680319] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b8568bf9-ac2f-4ff0-b515-b5e5351ae1d8 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.806926] env[62692]: DEBUG oslo_vmware.rw_handles [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526b55f8-d151-44c8-d576-e48b1c8fdeac/disk-0.vmdk. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1396.807923] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df99e1ba-e8e6-4dc3-9c3a-3f41d69aa0c1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.814434] env[62692]: DEBUG oslo_vmware.rw_handles [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526b55f8-d151-44c8-d576-e48b1c8fdeac/disk-0.vmdk is in state: ready. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1396.814633] env[62692]: ERROR oslo_vmware.rw_handles [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526b55f8-d151-44c8-d576-e48b1c8fdeac/disk-0.vmdk due to incomplete transfer. [ 1396.814864] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5649b488-3b69-4b3f-bb00-916336d5e476 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.822786] env[62692]: DEBUG oslo_vmware.rw_handles [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526b55f8-d151-44c8-d576-e48b1c8fdeac/disk-0.vmdk. {{(pid=62692) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1396.822997] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Uploaded image 8900c02f-8bce-4dc2-b362-b565762af1fd to the Glance image server {{(pid=62692) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1396.825454] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Destroying the VM {{(pid=62692) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1396.825693] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-71475bd4-79a2-44fb-b608-dc7a889b22b7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.832316] env[62692]: DEBUG oslo_vmware.api [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1396.832316] env[62692]: value = "task-1142238" [ 1396.832316] env[62692]: _type = "Task" [ 1396.832316] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1396.840092] env[62692]: DEBUG oslo_vmware.api [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142238, 'name': Destroy_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1396.856699] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "e4f1850a-5310-4ee7-a3b7-adbba64f5a83" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1396.856979] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "e4f1850a-5310-4ee7-a3b7-adbba64f5a83" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1396.857213] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "e4f1850a-5310-4ee7-a3b7-adbba64f5a83-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1396.857408] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "e4f1850a-5310-4ee7-a3b7-adbba64f5a83-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1396.857585] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "e4f1850a-5310-4ee7-a3b7-adbba64f5a83-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1396.859830] env[62692]: INFO nova.compute.manager [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Terminating instance [ 1396.861606] env[62692]: DEBUG nova.compute.manager [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1396.861812] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1396.862625] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2e78293-85dc-45a0-8d6b-136dad09efdc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.869284] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1396.869507] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-97d5465f-7e7b-496c-945b-922cd8021c75 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.875358] env[62692]: DEBUG oslo_vmware.api [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1396.875358] env[62692]: value = "task-1142239" [ 1396.875358] env[62692]: _type = "Task" [ 1396.875358] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1396.883072] env[62692]: DEBUG oslo_vmware.api [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142239, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1397.342234] env[62692]: DEBUG oslo_vmware.api [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142238, 'name': Destroy_Task, 'duration_secs': 0.360833} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1397.342439] env[62692]: INFO nova.virt.vmwareapi.vm_util [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Destroyed the VM [ 1397.342727] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Deleting Snapshot of the VM instance {{(pid=62692) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1397.342980] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4dfe28fe-8545-4a44-8c1d-d819ae2ef0e2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1397.349273] env[62692]: DEBUG oslo_vmware.api [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1397.349273] env[62692]: value = "task-1142240" [ 1397.349273] env[62692]: _type = "Task" [ 1397.349273] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1397.357636] env[62692]: DEBUG oslo_vmware.api [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142240, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1397.383717] env[62692]: DEBUG oslo_vmware.api [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142239, 'name': PowerOffVM_Task, 'duration_secs': 0.24124} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1397.383960] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1397.384154] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1397.384398] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-99642632-188a-42d6-8255-e9684d89b32c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1397.444098] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1397.444441] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Deleting contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1397.444699] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Deleting the datastore file [datastore1] e4f1850a-5310-4ee7-a3b7-adbba64f5a83 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1397.444980] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a96ac232-9aa4-47b0-a99e-11eb926f2df9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1397.451626] env[62692]: DEBUG oslo_vmware.api [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for the task: (returnval){ [ 1397.451626] env[62692]: value = "task-1142242" [ 1397.451626] env[62692]: _type = "Task" [ 1397.451626] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1397.459111] env[62692]: DEBUG oslo_vmware.api [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142242, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1397.858946] env[62692]: DEBUG oslo_vmware.api [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142240, 'name': RemoveSnapshot_Task, 'duration_secs': 0.376523} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1397.859253] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Deleted Snapshot of the VM instance {{(pid=62692) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1397.859527] env[62692]: DEBUG nova.compute.manager [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1397.860301] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39e436a4-b259-4a14-982a-f2927422659f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1397.965521] env[62692]: DEBUG oslo_vmware.api [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Task: {'id': task-1142242, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187906} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1397.965900] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1397.966197] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Deleted contents of the VM from datastore datastore1 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1397.966469] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1397.966735] env[62692]: INFO nova.compute.manager [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1397.967107] env[62692]: DEBUG oslo.service.loopingcall [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1397.967378] env[62692]: DEBUG nova.compute.manager [-] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1397.967518] env[62692]: DEBUG nova.network.neutron [-] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1398.372090] env[62692]: INFO nova.compute.manager [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Shelve offloading [ 1398.375227] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1398.375681] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f8d816c8-cd0a-4b4b-b84e-3f7d7181f067 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.382925] env[62692]: DEBUG oslo_vmware.api [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1398.382925] env[62692]: value = "task-1142243" [ 1398.382925] env[62692]: _type = "Task" [ 1398.382925] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1398.391276] env[62692]: DEBUG oslo_vmware.api [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142243, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1398.396244] env[62692]: DEBUG nova.compute.manager [req-712cf5a2-6b7a-41f2-bcef-2b527cc4e51a req-869beb22-fef6-48d0-827d-3e66b6e2ac63 service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Received event network-vif-deleted-1863d5f5-d859-4e5b-9d7f-b6e0df952e3d {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1398.396443] env[62692]: INFO nova.compute.manager [req-712cf5a2-6b7a-41f2-bcef-2b527cc4e51a req-869beb22-fef6-48d0-827d-3e66b6e2ac63 service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Neutron deleted interface 1863d5f5-d859-4e5b-9d7f-b6e0df952e3d; detaching it from the instance and deleting it from the info cache [ 1398.396620] env[62692]: DEBUG nova.network.neutron [req-712cf5a2-6b7a-41f2-bcef-2b527cc4e51a req-869beb22-fef6-48d0-827d-3e66b6e2ac63 service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1398.878153] env[62692]: DEBUG nova.network.neutron [-] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1398.894329] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] VM already powered off {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1398.894572] env[62692]: DEBUG nova.compute.manager [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1398.895363] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df117066-96de-490f-b97b-ef0492c591f4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.898721] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5315020f-81ec-4d8e-a989-48ca13c7b39f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.902758] env[62692]: DEBUG oslo_concurrency.lockutils [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "refresh_cache-6e5b2bef-8bdf-402e-b4de-87a54c6d221b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1398.902925] env[62692]: DEBUG oslo_concurrency.lockutils [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquired lock "refresh_cache-6e5b2bef-8bdf-402e-b4de-87a54c6d221b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1398.903117] env[62692]: DEBUG nova.network.neutron [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1398.910569] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ade2da1a-9ed9-4d06-9951-b47d3df38a59 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.934388] env[62692]: DEBUG nova.compute.manager [req-712cf5a2-6b7a-41f2-bcef-2b527cc4e51a req-869beb22-fef6-48d0-827d-3e66b6e2ac63 service nova] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Detach interface failed, port_id=1863d5f5-d859-4e5b-9d7f-b6e0df952e3d, reason: Instance e4f1850a-5310-4ee7-a3b7-adbba64f5a83 could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1399.382111] env[62692]: INFO nova.compute.manager [-] [instance: e4f1850a-5310-4ee7-a3b7-adbba64f5a83] Took 1.41 seconds to deallocate network for instance. [ 1399.602845] env[62692]: DEBUG nova.network.neutron [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Updating instance_info_cache with network_info: [{"id": "ef2af1cd-8e90-45c8-9f42-4fd1197a43dc", "address": "fa:16:3e:5a:d0:c9", "network": {"id": "77695f27-3027-4a92-8b9b-0a90cdc53f62", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-337681783-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697b4281872f4547ad3d051e644b95ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef2af1cd-8e", "ovs_interfaceid": "ef2af1cd-8e90-45c8-9f42-4fd1197a43dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1399.887902] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1399.888288] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1399.888430] env[62692]: DEBUG nova.objects.instance [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lazy-loading 'resources' on Instance uuid e4f1850a-5310-4ee7-a3b7-adbba64f5a83 {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1400.106154] env[62692]: DEBUG oslo_concurrency.lockutils [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Releasing lock "refresh_cache-6e5b2bef-8bdf-402e-b4de-87a54c6d221b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1400.430524] env[62692]: DEBUG nova.compute.manager [req-83ec7113-4177-43df-a928-a059944a3beb req-d1b2556b-b18e-4763-85a9-37586df82959 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Received event network-vif-unplugged-ef2af1cd-8e90-45c8-9f42-4fd1197a43dc {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1400.430758] env[62692]: DEBUG oslo_concurrency.lockutils [req-83ec7113-4177-43df-a928-a059944a3beb req-d1b2556b-b18e-4763-85a9-37586df82959 service nova] Acquiring lock "6e5b2bef-8bdf-402e-b4de-87a54c6d221b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1400.430976] env[62692]: DEBUG oslo_concurrency.lockutils [req-83ec7113-4177-43df-a928-a059944a3beb req-d1b2556b-b18e-4763-85a9-37586df82959 service nova] Lock "6e5b2bef-8bdf-402e-b4de-87a54c6d221b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1400.431193] env[62692]: DEBUG oslo_concurrency.lockutils [req-83ec7113-4177-43df-a928-a059944a3beb req-d1b2556b-b18e-4763-85a9-37586df82959 service nova] Lock "6e5b2bef-8bdf-402e-b4de-87a54c6d221b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1400.431319] env[62692]: DEBUG nova.compute.manager [req-83ec7113-4177-43df-a928-a059944a3beb req-d1b2556b-b18e-4763-85a9-37586df82959 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] No waiting events found dispatching network-vif-unplugged-ef2af1cd-8e90-45c8-9f42-4fd1197a43dc {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1400.431493] env[62692]: WARNING nova.compute.manager [req-83ec7113-4177-43df-a928-a059944a3beb req-d1b2556b-b18e-4763-85a9-37586df82959 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Received unexpected event network-vif-unplugged-ef2af1cd-8e90-45c8-9f42-4fd1197a43dc for instance with vm_state shelved and task_state shelving_offloading. [ 1400.436973] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4029706-2c6d-4664-814a-74e8aac0104f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.441682] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1400.442638] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b2a392d-e39c-4cf5-9784-7decd53aeca7 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.451564] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90401b1c-24c2-46a3-81e5-98107bee656f {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.454797] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1400.455014] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-682ece21-fae6-4777-8d27-eb329f436ddb {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.483688] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60331d3e-11c3-42ad-807d-91f468ae68d3 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.490464] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-757a4ec7-f948-4b78-9364-4c7b94299abd {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.504402] env[62692]: DEBUG nova.compute.provider_tree [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1400.522602] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1400.522966] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1400.523209] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Deleting the datastore file [datastore2] 6e5b2bef-8bdf-402e-b4de-87a54c6d221b {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1400.523458] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b646a776-0b11-404b-8f0b-850cadde49b0 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.529270] env[62692]: DEBUG oslo_vmware.api [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1400.529270] env[62692]: value = "task-1142245" [ 1400.529270] env[62692]: _type = "Task" [ 1400.529270] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1400.536865] env[62692]: DEBUG oslo_vmware.api [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142245, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1401.007715] env[62692]: DEBUG nova.scheduler.client.report [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1401.042925] env[62692]: DEBUG oslo_vmware.api [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142245, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145049} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1401.043155] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1401.043330] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1401.043507] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1401.064475] env[62692]: INFO nova.scheduler.client.report [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Deleted allocations for instance 6e5b2bef-8bdf-402e-b4de-87a54c6d221b [ 1401.513149] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.625s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1401.530534] env[62692]: INFO nova.scheduler.client.report [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Deleted allocations for instance e4f1850a-5310-4ee7-a3b7-adbba64f5a83 [ 1401.569058] env[62692]: DEBUG oslo_concurrency.lockutils [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1401.569409] env[62692]: DEBUG oslo_concurrency.lockutils [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1401.569572] env[62692]: DEBUG nova.objects.instance [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lazy-loading 'resources' on Instance uuid 6e5b2bef-8bdf-402e-b4de-87a54c6d221b {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1402.039420] env[62692]: DEBUG oslo_concurrency.lockutils [None req-6c56c995-69b8-4fd4-94c2-3b6c13fa4d39 tempest-AttachVolumeShelveTestJSON-1849164656 tempest-AttachVolumeShelveTestJSON-1849164656-project-member] Lock "e4f1850a-5310-4ee7-a3b7-adbba64f5a83" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.182s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1402.073084] env[62692]: DEBUG nova.objects.instance [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lazy-loading 'numa_topology' on Instance uuid 6e5b2bef-8bdf-402e-b4de-87a54c6d221b {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1402.459424] env[62692]: DEBUG nova.compute.manager [req-97394f58-5d41-4b92-bfa4-8e262ac070c5 req-9ae6f91e-a500-4791-acdb-1590f0e905f5 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Received event network-changed-ef2af1cd-8e90-45c8-9f42-4fd1197a43dc {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1402.459586] env[62692]: DEBUG nova.compute.manager [req-97394f58-5d41-4b92-bfa4-8e262ac070c5 req-9ae6f91e-a500-4791-acdb-1590f0e905f5 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Refreshing instance network info cache due to event network-changed-ef2af1cd-8e90-45c8-9f42-4fd1197a43dc. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1402.459803] env[62692]: DEBUG oslo_concurrency.lockutils [req-97394f58-5d41-4b92-bfa4-8e262ac070c5 req-9ae6f91e-a500-4791-acdb-1590f0e905f5 service nova] Acquiring lock "refresh_cache-6e5b2bef-8bdf-402e-b4de-87a54c6d221b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1402.459953] env[62692]: DEBUG oslo_concurrency.lockutils [req-97394f58-5d41-4b92-bfa4-8e262ac070c5 req-9ae6f91e-a500-4791-acdb-1590f0e905f5 service nova] Acquired lock "refresh_cache-6e5b2bef-8bdf-402e-b4de-87a54c6d221b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1402.460137] env[62692]: DEBUG nova.network.neutron [req-97394f58-5d41-4b92-bfa4-8e262ac070c5 req-9ae6f91e-a500-4791-acdb-1590f0e905f5 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Refreshing network info cache for port ef2af1cd-8e90-45c8-9f42-4fd1197a43dc {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1402.575680] env[62692]: DEBUG nova.objects.base [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Object Instance<6e5b2bef-8bdf-402e-b4de-87a54c6d221b> lazy-loaded attributes: resources,numa_topology {{(pid=62692) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1402.606586] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf744d09-fb7e-403e-97b7-5175ecdfc77c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1402.614249] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f042b88c-8850-4b8b-a8fa-257620499e4b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1402.644718] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7729b5d0-a446-4398-81e1-465597cab8bc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1402.651469] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af976dca-3689-40c4-be45-f8130eeddd4c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1402.664027] env[62692]: DEBUG nova.compute.provider_tree [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1402.680254] env[62692]: DEBUG oslo_concurrency.lockutils [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "6e5b2bef-8bdf-402e-b4de-87a54c6d221b" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1403.165606] env[62692]: DEBUG nova.network.neutron [req-97394f58-5d41-4b92-bfa4-8e262ac070c5 req-9ae6f91e-a500-4791-acdb-1590f0e905f5 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Updated VIF entry in instance network info cache for port ef2af1cd-8e90-45c8-9f42-4fd1197a43dc. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1403.166017] env[62692]: DEBUG nova.network.neutron [req-97394f58-5d41-4b92-bfa4-8e262ac070c5 req-9ae6f91e-a500-4791-acdb-1590f0e905f5 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Updating instance_info_cache with network_info: [{"id": "ef2af1cd-8e90-45c8-9f42-4fd1197a43dc", "address": "fa:16:3e:5a:d0:c9", "network": {"id": "77695f27-3027-4a92-8b9b-0a90cdc53f62", "bridge": null, "label": "tempest-ServerActionsTestOtherB-337681783-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697b4281872f4547ad3d051e644b95ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapef2af1cd-8e", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1403.167741] env[62692]: DEBUG nova.scheduler.client.report [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1403.671435] env[62692]: DEBUG oslo_concurrency.lockutils [req-97394f58-5d41-4b92-bfa4-8e262ac070c5 req-9ae6f91e-a500-4791-acdb-1590f0e905f5 service nova] Releasing lock "refresh_cache-6e5b2bef-8bdf-402e-b4de-87a54c6d221b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1403.672193] env[62692]: DEBUG oslo_concurrency.lockutils [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.103s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1404.180102] env[62692]: DEBUG oslo_concurrency.lockutils [None req-338291be-e386-4709-b71f-2c4659d12c82 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "6e5b2bef-8bdf-402e-b4de-87a54c6d221b" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 19.263s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1404.180600] env[62692]: DEBUG oslo_concurrency.lockutils [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "6e5b2bef-8bdf-402e-b4de-87a54c6d221b" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.500s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1404.180792] env[62692]: INFO nova.compute.manager [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Unshelving [ 1405.206302] env[62692]: DEBUG oslo_concurrency.lockutils [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1405.206599] env[62692]: DEBUG oslo_concurrency.lockutils [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1405.206830] env[62692]: DEBUG nova.objects.instance [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lazy-loading 'pci_requests' on Instance uuid 6e5b2bef-8bdf-402e-b4de-87a54c6d221b {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1405.710317] env[62692]: DEBUG nova.objects.instance [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lazy-loading 'numa_topology' on Instance uuid 6e5b2bef-8bdf-402e-b4de-87a54c6d221b {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1406.213276] env[62692]: INFO nova.compute.claims [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1407.257512] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2943cb3-4c6e-40c8-bd1c-1b10484a4d94 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.267160] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0df74a8-b120-413a-a130-20ea62400b8c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.301098] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61b63708-854e-4ed8-b988-e1e43f9399a4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.312031] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb89c3ef-0069-4db2-8e02-512e43f8b8e1 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.321043] env[62692]: DEBUG nova.compute.provider_tree [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Inventory has not changed in ProviderTree for provider: 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1407.824256] env[62692]: DEBUG nova.scheduler.client.report [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Inventory has not changed for provider 470758a8-5efa-4875-aad5-f512a727752f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1408.329670] env[62692]: DEBUG oslo_concurrency.lockutils [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.123s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1408.366613] env[62692]: INFO nova.network.neutron [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Updating port ef2af1cd-8e90-45c8-9f42-4fd1197a43dc with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1409.749934] env[62692]: DEBUG nova.compute.manager [req-b8562d82-64af-4c19-a9f6-050706d5274d req-92b10e53-7adf-4a4a-99ca-2db608b37d46 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Received event network-vif-plugged-ef2af1cd-8e90-45c8-9f42-4fd1197a43dc {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1409.750192] env[62692]: DEBUG oslo_concurrency.lockutils [req-b8562d82-64af-4c19-a9f6-050706d5274d req-92b10e53-7adf-4a4a-99ca-2db608b37d46 service nova] Acquiring lock "6e5b2bef-8bdf-402e-b4de-87a54c6d221b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1409.750395] env[62692]: DEBUG oslo_concurrency.lockutils [req-b8562d82-64af-4c19-a9f6-050706d5274d req-92b10e53-7adf-4a4a-99ca-2db608b37d46 service nova] Lock "6e5b2bef-8bdf-402e-b4de-87a54c6d221b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1409.750567] env[62692]: DEBUG oslo_concurrency.lockutils [req-b8562d82-64af-4c19-a9f6-050706d5274d req-92b10e53-7adf-4a4a-99ca-2db608b37d46 service nova] Lock "6e5b2bef-8bdf-402e-b4de-87a54c6d221b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1409.750738] env[62692]: DEBUG nova.compute.manager [req-b8562d82-64af-4c19-a9f6-050706d5274d req-92b10e53-7adf-4a4a-99ca-2db608b37d46 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] No waiting events found dispatching network-vif-plugged-ef2af1cd-8e90-45c8-9f42-4fd1197a43dc {{(pid=62692) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1409.750912] env[62692]: WARNING nova.compute.manager [req-b8562d82-64af-4c19-a9f6-050706d5274d req-92b10e53-7adf-4a4a-99ca-2db608b37d46 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Received unexpected event network-vif-plugged-ef2af1cd-8e90-45c8-9f42-4fd1197a43dc for instance with vm_state shelved_offloaded and task_state spawning. [ 1409.830232] env[62692]: DEBUG oslo_concurrency.lockutils [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "refresh_cache-6e5b2bef-8bdf-402e-b4de-87a54c6d221b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1409.830395] env[62692]: DEBUG oslo_concurrency.lockutils [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquired lock "refresh_cache-6e5b2bef-8bdf-402e-b4de-87a54c6d221b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1409.830606] env[62692]: DEBUG nova.network.neutron [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Building network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1410.527119] env[62692]: DEBUG nova.network.neutron [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Updating instance_info_cache with network_info: [{"id": "ef2af1cd-8e90-45c8-9f42-4fd1197a43dc", "address": "fa:16:3e:5a:d0:c9", "network": {"id": "77695f27-3027-4a92-8b9b-0a90cdc53f62", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-337681783-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697b4281872f4547ad3d051e644b95ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef2af1cd-8e", "ovs_interfaceid": "ef2af1cd-8e90-45c8-9f42-4fd1197a43dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1410.627476] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1410.627640] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1411.029515] env[62692]: DEBUG oslo_concurrency.lockutils [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Releasing lock "refresh_cache-6e5b2bef-8bdf-402e-b4de-87a54c6d221b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1411.058896] env[62692]: DEBUG nova.virt.hardware [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-05T20:28:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='32029ec8b340a848f23a2ca51bdebc3b',container_format='bare',created_at=2024-10-05T20:42:49Z,direct_url=,disk_format='vmdk',id=8900c02f-8bce-4dc2-b362-b565762af1fd,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-1540345706-shelved',owner='697b4281872f4547ad3d051e644b95ce',properties=ImageMetaProps,protected=,size=31669760,status='active',tags=,updated_at=2024-10-05T20:43:02Z,virtual_size=,visibility=), allow threads: False {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1411.059175] env[62692]: DEBUG nova.virt.hardware [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Flavor limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1411.059345] env[62692]: DEBUG nova.virt.hardware [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Image limits 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1411.059536] env[62692]: DEBUG nova.virt.hardware [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Flavor pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1411.059692] env[62692]: DEBUG nova.virt.hardware [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Image pref 0:0:0 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1411.059847] env[62692]: DEBUG nova.virt.hardware [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62692) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1411.060069] env[62692]: DEBUG nova.virt.hardware [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1411.060242] env[62692]: DEBUG nova.virt.hardware [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1411.060414] env[62692]: DEBUG nova.virt.hardware [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Got 1 possible topologies {{(pid=62692) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1411.060580] env[62692]: DEBUG nova.virt.hardware [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1411.060803] env[62692]: DEBUG nova.virt.hardware [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62692) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1411.061661] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1d3d656-154a-4a9b-b9be-65e7d8d4238c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1411.069135] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0754ac4-2aff-490e-971f-1a5ea8dc62df {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1411.081717] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5a:d0:c9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8e7f6f41-f4eb-4832-a390-730fca1cf717', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ef2af1cd-8e90-45c8-9f42-4fd1197a43dc', 'vif_model': 'vmxnet3'}] {{(pid=62692) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1411.089010] env[62692]: DEBUG oslo.service.loopingcall [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1411.089247] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Creating VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1411.089448] env[62692]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b60b954d-7f89-45ba-9d6d-d1f3e620a74c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1411.108509] env[62692]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1411.108509] env[62692]: value = "task-1142247" [ 1411.108509] env[62692]: _type = "Task" [ 1411.108509] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1411.115954] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1142247, 'name': CreateVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1411.132224] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1411.132224] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Starting heal instance info cache {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1411.618693] env[62692]: DEBUG oslo_vmware.api [-] Task: {'id': task-1142247, 'name': CreateVM_Task, 'duration_secs': 0.318826} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1411.618860] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Created VM on the ESX host {{(pid=62692) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1411.625861] env[62692]: DEBUG oslo_concurrency.lockutils [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8900c02f-8bce-4dc2-b362-b565762af1fd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1411.626058] env[62692]: DEBUG oslo_concurrency.lockutils [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8900c02f-8bce-4dc2-b362-b565762af1fd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1411.626454] env[62692]: DEBUG oslo_concurrency.lockutils [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8900c02f-8bce-4dc2-b362-b565762af1fd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1411.626698] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc4e3b73-2632-49da-bd4b-46ddaf8de0c5 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1411.630640] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1411.630640] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52309c92-4524-b0b0-bab3-0be36595176e" [ 1411.630640] env[62692]: _type = "Task" [ 1411.630640] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1411.635927] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "refresh_cache-6e5b2bef-8bdf-402e-b4de-87a54c6d221b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1411.636068] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquired lock "refresh_cache-6e5b2bef-8bdf-402e-b4de-87a54c6d221b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1411.636213] env[62692]: DEBUG nova.network.neutron [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Forcefully refreshing network info cache for instance {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1411.639782] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]52309c92-4524-b0b0-bab3-0be36595176e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1411.772900] env[62692]: DEBUG nova.compute.manager [req-0a1e38da-e2dd-4042-915e-071bb50cb260 req-9a6f21a6-4371-4040-bd9b-da4bb01cf6f4 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Received event network-changed-ef2af1cd-8e90-45c8-9f42-4fd1197a43dc {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1411.773064] env[62692]: DEBUG nova.compute.manager [req-0a1e38da-e2dd-4042-915e-071bb50cb260 req-9a6f21a6-4371-4040-bd9b-da4bb01cf6f4 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Refreshing instance network info cache due to event network-changed-ef2af1cd-8e90-45c8-9f42-4fd1197a43dc. {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1411.773314] env[62692]: DEBUG oslo_concurrency.lockutils [req-0a1e38da-e2dd-4042-915e-071bb50cb260 req-9a6f21a6-4371-4040-bd9b-da4bb01cf6f4 service nova] Acquiring lock "refresh_cache-6e5b2bef-8bdf-402e-b4de-87a54c6d221b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1412.144024] env[62692]: DEBUG oslo_concurrency.lockutils [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8900c02f-8bce-4dc2-b362-b565762af1fd" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1412.144024] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Processing image 8900c02f-8bce-4dc2-b362-b565762af1fd {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1412.144024] env[62692]: DEBUG oslo_concurrency.lockutils [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8900c02f-8bce-4dc2-b362-b565762af1fd/8900c02f-8bce-4dc2-b362-b565762af1fd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1412.144024] env[62692]: DEBUG oslo_concurrency.lockutils [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8900c02f-8bce-4dc2-b362-b565762af1fd/8900c02f-8bce-4dc2-b362-b565762af1fd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1412.144593] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1412.145315] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1488a04d-00e8-41c0-9ec1-2c3a6818619d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1412.163454] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1412.163454] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62692) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1412.163454] env[62692]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2abe4dec-a604-4d2b-9a0e-59c1bdf51cad {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1412.168603] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1412.168603] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]521bbbff-289d-2d1b-871e-678c2cabe45a" [ 1412.168603] env[62692]: _type = "Task" [ 1412.168603] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1412.177284] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': session[52caca42-f181-c60e-9702-6892b2d97966]521bbbff-289d-2d1b-871e-678c2cabe45a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1412.679641] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Preparing fetch location {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1412.679896] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Fetch image to [datastore2] OSTACK_IMG_20e5254f-da6f-41c4-a18a-83938134f604/OSTACK_IMG_20e5254f-da6f-41c4-a18a-83938134f604.vmdk {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1412.680095] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Downloading stream optimized image 8900c02f-8bce-4dc2-b362-b565762af1fd to [datastore2] OSTACK_IMG_20e5254f-da6f-41c4-a18a-83938134f604/OSTACK_IMG_20e5254f-da6f-41c4-a18a-83938134f604.vmdk on the data store datastore2 as vApp {{(pid=62692) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1412.680273] env[62692]: DEBUG nova.virt.vmwareapi.images [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Downloading image file data 8900c02f-8bce-4dc2-b362-b565762af1fd to the ESX as VM named 'OSTACK_IMG_20e5254f-da6f-41c4-a18a-83938134f604' {{(pid=62692) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1412.746226] env[62692]: DEBUG oslo_vmware.rw_handles [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1412.746226] env[62692]: value = "resgroup-9" [ 1412.746226] env[62692]: _type = "ResourcePool" [ 1412.746226] env[62692]: }. {{(pid=62692) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1412.746512] env[62692]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-3e36a29b-cc39-4c8b-bbcb-0c18b4e561f4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1412.771169] env[62692]: DEBUG oslo_vmware.rw_handles [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lease: (returnval){ [ 1412.771169] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52e21212-a98d-5bcb-5aca-9ee4e4fb33f0" [ 1412.771169] env[62692]: _type = "HttpNfcLease" [ 1412.771169] env[62692]: } obtained for vApp import into resource pool (val){ [ 1412.771169] env[62692]: value = "resgroup-9" [ 1412.771169] env[62692]: _type = "ResourcePool" [ 1412.771169] env[62692]: }. {{(pid=62692) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1412.771562] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the lease: (returnval){ [ 1412.771562] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52e21212-a98d-5bcb-5aca-9ee4e4fb33f0" [ 1412.771562] env[62692]: _type = "HttpNfcLease" [ 1412.771562] env[62692]: } to be ready. {{(pid=62692) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1412.777262] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1412.777262] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52e21212-a98d-5bcb-5aca-9ee4e4fb33f0" [ 1412.777262] env[62692]: _type = "HttpNfcLease" [ 1412.777262] env[62692]: } is initializing. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1412.853979] env[62692]: DEBUG nova.network.neutron [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Updating instance_info_cache with network_info: [{"id": "ef2af1cd-8e90-45c8-9f42-4fd1197a43dc", "address": "fa:16:3e:5a:d0:c9", "network": {"id": "77695f27-3027-4a92-8b9b-0a90cdc53f62", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-337681783-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697b4281872f4547ad3d051e644b95ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef2af1cd-8e", "ovs_interfaceid": "ef2af1cd-8e90-45c8-9f42-4fd1197a43dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1413.279655] env[62692]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1413.279655] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52e21212-a98d-5bcb-5aca-9ee4e4fb33f0" [ 1413.279655] env[62692]: _type = "HttpNfcLease" [ 1413.279655] env[62692]: } is ready. {{(pid=62692) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1413.280532] env[62692]: DEBUG oslo_vmware.rw_handles [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1413.280532] env[62692]: value = "session[52caca42-f181-c60e-9702-6892b2d97966]52e21212-a98d-5bcb-5aca-9ee4e4fb33f0" [ 1413.280532] env[62692]: _type = "HttpNfcLease" [ 1413.280532] env[62692]: }. {{(pid=62692) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1413.280776] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ba8c45c-404a-48c4-b77a-8fbf6029c23d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.288175] env[62692]: DEBUG oslo_vmware.rw_handles [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52319864-bb53-86ae-526d-d508dc9d8b2d/disk-0.vmdk from lease info. {{(pid=62692) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1413.288382] env[62692]: DEBUG oslo_vmware.rw_handles [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Creating HTTP connection to write to file with size = 31669760 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52319864-bb53-86ae-526d-d508dc9d8b2d/disk-0.vmdk. {{(pid=62692) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1413.350795] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8a21c07f-523c-41ce-8477-2ad3a4ee2be2 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.357107] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Releasing lock "refresh_cache-6e5b2bef-8bdf-402e-b4de-87a54c6d221b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1413.357399] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Updated the network info_cache for instance {{(pid=62692) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1413.358853] env[62692]: DEBUG oslo_concurrency.lockutils [req-0a1e38da-e2dd-4042-915e-071bb50cb260 req-9a6f21a6-4371-4040-bd9b-da4bb01cf6f4 service nova] Acquired lock "refresh_cache-6e5b2bef-8bdf-402e-b4de-87a54c6d221b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1413.359171] env[62692]: DEBUG nova.network.neutron [req-0a1e38da-e2dd-4042-915e-071bb50cb260 req-9a6f21a6-4371-4040-bd9b-da4bb01cf6f4 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Refreshing network info cache for port ef2af1cd-8e90-45c8-9f42-4fd1197a43dc {{(pid=62692) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1413.360460] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1413.360778] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1413.361856] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1413.361856] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1413.361985] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1413.362091] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1413.362247] env[62692]: DEBUG nova.compute.manager [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62692) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1413.362418] env[62692]: DEBUG oslo_service.periodic_task [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Running periodic task ComputeManager.update_available_resource {{(pid=62692) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1413.865944] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1413.866203] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1413.866370] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1413.866526] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62692) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1413.867829] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66e48d96-4a8a-4137-bd33-ed0e1d77f923 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.876553] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39d27397-700e-4899-8bb5-5dde80086b1d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.894944] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e23afcde-a6a4-43ca-a8b7-41940dc8e7ec {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.906144] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b3e0a11-7d17-4536-82d8-32e30cecb354 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.935341] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181283MB free_disk=167GB free_vcpus=48 pci_devices=None {{(pid=62692) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1413.935526] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1413.935772] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1414.119595] env[62692]: DEBUG nova.network.neutron [req-0a1e38da-e2dd-4042-915e-071bb50cb260 req-9a6f21a6-4371-4040-bd9b-da4bb01cf6f4 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Updated VIF entry in instance network info cache for port ef2af1cd-8e90-45c8-9f42-4fd1197a43dc. {{(pid=62692) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1414.120146] env[62692]: DEBUG nova.network.neutron [req-0a1e38da-e2dd-4042-915e-071bb50cb260 req-9a6f21a6-4371-4040-bd9b-da4bb01cf6f4 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Updating instance_info_cache with network_info: [{"id": "ef2af1cd-8e90-45c8-9f42-4fd1197a43dc", "address": "fa:16:3e:5a:d0:c9", "network": {"id": "77695f27-3027-4a92-8b9b-0a90cdc53f62", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-337681783-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697b4281872f4547ad3d051e644b95ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef2af1cd-8e", "ovs_interfaceid": "ef2af1cd-8e90-45c8-9f42-4fd1197a43dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1414.494878] env[62692]: DEBUG oslo_vmware.rw_handles [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Completed reading data from the image iterator. {{(pid=62692) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1414.495286] env[62692]: DEBUG oslo_vmware.rw_handles [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52319864-bb53-86ae-526d-d508dc9d8b2d/disk-0.vmdk. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1414.496065] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f2377f3-6ea8-45de-9f06-9a25ebcd7b4a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1414.502729] env[62692]: DEBUG oslo_vmware.rw_handles [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52319864-bb53-86ae-526d-d508dc9d8b2d/disk-0.vmdk is in state: ready. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1414.502891] env[62692]: DEBUG oslo_vmware.rw_handles [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52319864-bb53-86ae-526d-d508dc9d8b2d/disk-0.vmdk. {{(pid=62692) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1414.503140] env[62692]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-88994456-85f6-425a-9e90-101a85cf5e36 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1414.622615] env[62692]: DEBUG oslo_concurrency.lockutils [req-0a1e38da-e2dd-4042-915e-071bb50cb260 req-9a6f21a6-4371-4040-bd9b-da4bb01cf6f4 service nova] Releasing lock "refresh_cache-6e5b2bef-8bdf-402e-b4de-87a54c6d221b" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1414.696257] env[62692]: DEBUG oslo_vmware.rw_handles [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52319864-bb53-86ae-526d-d508dc9d8b2d/disk-0.vmdk. {{(pid=62692) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1414.696544] env[62692]: INFO nova.virt.vmwareapi.images [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Downloaded image file data 8900c02f-8bce-4dc2-b362-b565762af1fd [ 1414.697397] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24a8a3df-4d1a-47ad-adca-bb00de8cee1a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1414.713657] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b3cb8e8b-610c-4a97-ac60-ddc04bb39f68 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1414.759581] env[62692]: INFO nova.virt.vmwareapi.images [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] The imported VM was unregistered [ 1414.762032] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Caching image {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1414.762276] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Creating directory with path [datastore2] devstack-image-cache_base/8900c02f-8bce-4dc2-b362-b565762af1fd {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1414.762540] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f73dd969-2c4f-453f-82b7-5850d1ac8d2e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1414.780589] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Created directory with path [datastore2] devstack-image-cache_base/8900c02f-8bce-4dc2-b362-b565762af1fd {{(pid=62692) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1414.780774] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_20e5254f-da6f-41c4-a18a-83938134f604/OSTACK_IMG_20e5254f-da6f-41c4-a18a-83938134f604.vmdk to [datastore2] devstack-image-cache_base/8900c02f-8bce-4dc2-b362-b565762af1fd/8900c02f-8bce-4dc2-b362-b565762af1fd.vmdk. {{(pid=62692) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1414.781016] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-a4507057-52cb-4e3d-89f5-6039573dff4a {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1414.787169] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1414.787169] env[62692]: value = "task-1142250" [ 1414.787169] env[62692]: _type = "Task" [ 1414.787169] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1414.794959] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142250, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1414.972009] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Instance 6e5b2bef-8bdf-402e-b4de-87a54c6d221b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62692) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1414.972259] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62692) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1414.972392] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62692) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1415.005789] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b8766ab-fb32-4c71-94d4-f9d5b8c56b1e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1415.015756] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58609801-9610-4576-9e57-045981ae783d {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1415.050548] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa1aa447-295e-4cb0-8ef2-21473f5d4f59 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1415.060712] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece3da19-ce6b-4892-8930-cc2818f7cb54 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1415.077372] env[62692]: DEBUG nova.compute.provider_tree [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1415.300483] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142250, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1415.606279] env[62692]: ERROR nova.scheduler.client.report [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] [req-340c585c-b39d-4746-a26f-b068b4bbe725] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 470758a8-5efa-4875-aad5-f512a727752f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-340c585c-b39d-4746-a26f-b068b4bbe725"}]} [ 1415.623801] env[62692]: DEBUG nova.scheduler.client.report [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Refreshing inventories for resource provider 470758a8-5efa-4875-aad5-f512a727752f {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1415.639174] env[62692]: DEBUG nova.scheduler.client.report [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Updating ProviderTree inventory for provider 470758a8-5efa-4875-aad5-f512a727752f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1415.639379] env[62692]: DEBUG nova.compute.provider_tree [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1415.651140] env[62692]: DEBUG nova.scheduler.client.report [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Refreshing aggregate associations for resource provider 470758a8-5efa-4875-aad5-f512a727752f, aggregates: None {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1415.673376] env[62692]: DEBUG nova.scheduler.client.report [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Refreshing trait associations for resource provider 470758a8-5efa-4875-aad5-f512a727752f, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=62692) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1415.700403] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05d3caf0-67d5-4a5d-82ab-2c024eda5b38 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1415.712195] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b39c0c7-3128-4ef1-826e-379caa07f355 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1415.746744] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6085403-1a0b-4ae0-964c-14fcb40c13f9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1415.757681] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03642f01-7f80-470d-a1ff-1a250ec700a6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1415.774992] env[62692]: DEBUG nova.compute.provider_tree [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1415.802568] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142250, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1416.301059] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142250, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1416.304988] env[62692]: DEBUG nova.scheduler.client.report [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Updated inventory for provider 470758a8-5efa-4875-aad5-f512a727752f with generation 172 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1416.305231] env[62692]: DEBUG nova.compute.provider_tree [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Updating resource provider 470758a8-5efa-4875-aad5-f512a727752f generation from 172 to 173 during operation: update_inventory {{(pid=62692) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1416.305386] env[62692]: DEBUG nova.compute.provider_tree [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1416.803128] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142250, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1416.810021] env[62692]: DEBUG nova.compute.resource_tracker [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62692) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1416.810261] env[62692]: DEBUG oslo_concurrency.lockutils [None req-620329cc-ea60-4c4d-9b3c-9a57d4cc5aeb None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.874s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1417.300976] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142250, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.214668} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1417.301331] env[62692]: INFO nova.virt.vmwareapi.ds_util [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_20e5254f-da6f-41c4-a18a-83938134f604/OSTACK_IMG_20e5254f-da6f-41c4-a18a-83938134f604.vmdk to [datastore2] devstack-image-cache_base/8900c02f-8bce-4dc2-b362-b565762af1fd/8900c02f-8bce-4dc2-b362-b565762af1fd.vmdk. [ 1417.301567] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Cleaning up location [datastore2] OSTACK_IMG_20e5254f-da6f-41c4-a18a-83938134f604 {{(pid=62692) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1417.301766] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_20e5254f-da6f-41c4-a18a-83938134f604 {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1417.302130] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-203aa226-67ab-459f-83df-98cd2b0e4a10 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1417.309091] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1417.309091] env[62692]: value = "task-1142251" [ 1417.309091] env[62692]: _type = "Task" [ 1417.309091] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1417.316709] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142251, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1417.818946] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142251, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.034779} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1417.819365] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1417.819409] env[62692]: DEBUG oslo_concurrency.lockutils [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8900c02f-8bce-4dc2-b362-b565762af1fd/8900c02f-8bce-4dc2-b362-b565762af1fd.vmdk" {{(pid=62692) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1417.819652] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8900c02f-8bce-4dc2-b362-b565762af1fd/8900c02f-8bce-4dc2-b362-b565762af1fd.vmdk to [datastore2] 6e5b2bef-8bdf-402e-b4de-87a54c6d221b/6e5b2bef-8bdf-402e-b4de-87a54c6d221b.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1417.819939] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c0a258ea-5b67-4e2d-9c0e-f5e81c5b5c37 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1417.827005] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1417.827005] env[62692]: value = "task-1142252" [ 1417.827005] env[62692]: _type = "Task" [ 1417.827005] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1417.834134] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142252, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1418.341326] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142252, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1418.840961] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142252, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1419.342356] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142252, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1419.842318] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142252, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1420.340542] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142252, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.206248} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1420.340829] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8900c02f-8bce-4dc2-b362-b565762af1fd/8900c02f-8bce-4dc2-b362-b565762af1fd.vmdk to [datastore2] 6e5b2bef-8bdf-402e-b4de-87a54c6d221b/6e5b2bef-8bdf-402e-b4de-87a54c6d221b.vmdk {{(pid=62692) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1420.341581] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a108bc9-5ed5-44a2-8070-1b952e941e5b {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.363941] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Reconfiguring VM instance instance-0000007a to attach disk [datastore2] 6e5b2bef-8bdf-402e-b4de-87a54c6d221b/6e5b2bef-8bdf-402e-b4de-87a54c6d221b.vmdk or device None with type streamOptimized {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1420.364262] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-054cdc2a-c156-481f-af59-e49454708e67 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.383822] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1420.383822] env[62692]: value = "task-1142253" [ 1420.383822] env[62692]: _type = "Task" [ 1420.383822] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1420.391280] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142253, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1420.893692] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142253, 'name': ReconfigVM_Task, 'duration_secs': 0.452672} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1420.894046] env[62692]: DEBUG nova.virt.vmwareapi.volumeops [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Reconfigured VM instance instance-0000007a to attach disk [datastore2] 6e5b2bef-8bdf-402e-b4de-87a54c6d221b/6e5b2bef-8bdf-402e-b4de-87a54c6d221b.vmdk or device None with type streamOptimized {{(pid=62692) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1420.894690] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f6971fe6-4190-4963-b15e-17c46e2c1f01 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.901039] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1420.901039] env[62692]: value = "task-1142254" [ 1420.901039] env[62692]: _type = "Task" [ 1420.901039] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1420.908498] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142254, 'name': Rename_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1421.411456] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142254, 'name': Rename_Task, 'duration_secs': 0.206555} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1421.411718] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Powering on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1421.412017] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ed0f6055-0dca-4fd8-b086-c436f50aebbc {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1421.418315] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1421.418315] env[62692]: value = "task-1142255" [ 1421.418315] env[62692]: _type = "Task" [ 1421.418315] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1421.425741] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142255, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1421.928102] env[62692]: DEBUG oslo_vmware.api [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142255, 'name': PowerOnVM_Task, 'duration_secs': 0.404839} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1421.928515] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Powered on the VM {{(pid=62692) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1422.028461] env[62692]: DEBUG nova.compute.manager [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Checking state {{(pid=62692) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1422.029406] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5322605-e7ef-484c-a585-34e64417d574 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1422.547187] env[62692]: DEBUG oslo_concurrency.lockutils [None req-63151aca-6586-430f-8ce6-25a4ff1e7566 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "6e5b2bef-8bdf-402e-b4de-87a54c6d221b" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 18.366s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1423.918102] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "6e5b2bef-8bdf-402e-b4de-87a54c6d221b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1423.918437] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "6e5b2bef-8bdf-402e-b4de-87a54c6d221b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1423.918651] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "6e5b2bef-8bdf-402e-b4de-87a54c6d221b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1423.918843] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "6e5b2bef-8bdf-402e-b4de-87a54c6d221b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1423.919037] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "6e5b2bef-8bdf-402e-b4de-87a54c6d221b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1423.921209] env[62692]: INFO nova.compute.manager [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Terminating instance [ 1423.922983] env[62692]: DEBUG nova.compute.manager [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Start destroying the instance on the hypervisor. {{(pid=62692) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1423.923199] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Destroying instance {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1423.924033] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab694552-9eb2-4de8-9395-4e0b02c3039c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.932131] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Powering off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1423.932354] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1b0cea8d-f5e1-44cb-bc56-05f233dd48b4 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.938097] env[62692]: DEBUG oslo_vmware.api [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1423.938097] env[62692]: value = "task-1142256" [ 1423.938097] env[62692]: _type = "Task" [ 1423.938097] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1423.945472] env[62692]: DEBUG oslo_vmware.api [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142256, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1424.447800] env[62692]: DEBUG oslo_vmware.api [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142256, 'name': PowerOffVM_Task, 'duration_secs': 0.190967} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1424.448100] env[62692]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Powered off the VM {{(pid=62692) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1424.448279] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Unregistering the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1424.448524] env[62692]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e77b873c-5199-45ef-87fe-0103286f8429 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.505514] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Unregistered the VM {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1424.505769] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Deleting contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1424.505946] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Deleting the datastore file [datastore2] 6e5b2bef-8bdf-402e-b4de-87a54c6d221b {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1424.506234] env[62692]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-20ea9ebd-c10f-4892-9209-fc19186b2339 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.511800] env[62692]: DEBUG oslo_vmware.api [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for the task: (returnval){ [ 1424.511800] env[62692]: value = "task-1142258" [ 1424.511800] env[62692]: _type = "Task" [ 1424.511800] env[62692]: } to complete. {{(pid=62692) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1424.520466] env[62692]: DEBUG oslo_vmware.api [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142258, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1425.021435] env[62692]: DEBUG oslo_vmware.api [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Task: {'id': task-1142258, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133563} completed successfully. {{(pid=62692) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1425.021840] env[62692]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Deleted the datastore file {{(pid=62692) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1425.021887] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Deleted contents of the VM from datastore datastore2 {{(pid=62692) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1425.022061] env[62692]: DEBUG nova.virt.vmwareapi.vmops [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Instance destroyed {{(pid=62692) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1425.022247] env[62692]: INFO nova.compute.manager [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1425.022497] env[62692]: DEBUG oslo.service.loopingcall [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62692) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1425.022691] env[62692]: DEBUG nova.compute.manager [-] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Deallocating network for instance {{(pid=62692) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1425.022803] env[62692]: DEBUG nova.network.neutron [-] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] deallocate_for_instance() {{(pid=62692) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1425.428399] env[62692]: DEBUG nova.compute.manager [req-cb56d21a-da93-4820-bd5d-6ada0dd906e8 req-5fb5d3e4-7bc0-4a6c-a41a-a330ced85522 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Received event network-vif-deleted-ef2af1cd-8e90-45c8-9f42-4fd1197a43dc {{(pid=62692) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1425.428634] env[62692]: INFO nova.compute.manager [req-cb56d21a-da93-4820-bd5d-6ada0dd906e8 req-5fb5d3e4-7bc0-4a6c-a41a-a330ced85522 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Neutron deleted interface ef2af1cd-8e90-45c8-9f42-4fd1197a43dc; detaching it from the instance and deleting it from the info cache [ 1425.428805] env[62692]: DEBUG nova.network.neutron [req-cb56d21a-da93-4820-bd5d-6ada0dd906e8 req-5fb5d3e4-7bc0-4a6c-a41a-a330ced85522 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1425.904397] env[62692]: DEBUG nova.network.neutron [-] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Updating instance_info_cache with network_info: [] {{(pid=62692) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1425.931611] env[62692]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-240a4241-f678-4378-bd58-0aa47c0d83e6 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1425.941792] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5ca88dc-3f62-4860-8047-3c8f31f04fb9 {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1425.966874] env[62692]: DEBUG nova.compute.manager [req-cb56d21a-da93-4820-bd5d-6ada0dd906e8 req-5fb5d3e4-7bc0-4a6c-a41a-a330ced85522 service nova] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Detach interface failed, port_id=ef2af1cd-8e90-45c8-9f42-4fd1197a43dc, reason: Instance 6e5b2bef-8bdf-402e-b4de-87a54c6d221b could not be found. {{(pid=62692) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1426.409749] env[62692]: INFO nova.compute.manager [-] [instance: 6e5b2bef-8bdf-402e-b4de-87a54c6d221b] Took 1.39 seconds to deallocate network for instance. [ 1426.916612] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1426.916837] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1426.917104] env[62692]: DEBUG nova.objects.instance [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lazy-loading 'resources' on Instance uuid 6e5b2bef-8bdf-402e-b4de-87a54c6d221b {{(pid=62692) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1427.457944] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b55407e-9077-4c1d-b4eb-77901837222e {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1427.465867] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94d448e3-460c-4e48-9829-cbf827acedba {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1427.496283] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2c91042-fa52-4510-8390-fa4b0da1af7c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1427.503681] env[62692]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b27722d2-a361-425b-9214-89a8708b547c {{(pid=62692) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1427.516803] env[62692]: DEBUG nova.compute.provider_tree [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1428.046064] env[62692]: DEBUG nova.scheduler.client.report [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Updated inventory for provider 470758a8-5efa-4875-aad5-f512a727752f with generation 173 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1428.046360] env[62692]: DEBUG nova.compute.provider_tree [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Updating resource provider 470758a8-5efa-4875-aad5-f512a727752f generation from 173 to 174 during operation: update_inventory {{(pid=62692) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1428.046551] env[62692]: DEBUG nova.compute.provider_tree [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Updating inventory in ProviderTree for provider 470758a8-5efa-4875-aad5-f512a727752f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 166, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62692) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1428.551757] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.635s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1428.571980] env[62692]: INFO nova.scheduler.client.report [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Deleted allocations for instance 6e5b2bef-8bdf-402e-b4de-87a54c6d221b [ 1429.081856] env[62692]: DEBUG oslo_concurrency.lockutils [None req-b5ee3ecf-9adc-4987-b400-6dd1c929e0f7 tempest-ServerActionsTestOtherB-1722036111 tempest-ServerActionsTestOtherB-1722036111-project-member] Lock "6e5b2bef-8bdf-402e-b4de-87a54c6d221b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.163s {{(pid=62692) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}